var/home/core/zuul-output/0000755000175000017500000000000015071704010014520 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071714463015501 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004456016715071714453017721 0ustar rootrootOct 09 10:05:12 crc systemd[1]: Starting Kubernetes Kubelet... Oct 09 10:05:12 crc restorecon[4726]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:12 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 10:05:13 crc restorecon[4726]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 09 10:05:14 crc kubenswrapper[4923]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 09 10:05:14 crc kubenswrapper[4923]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 09 10:05:14 crc kubenswrapper[4923]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 09 10:05:14 crc kubenswrapper[4923]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 09 10:05:14 crc kubenswrapper[4923]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 09 10:05:14 crc kubenswrapper[4923]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.376179 4923 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380515 4923 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380542 4923 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380550 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380557 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380564 4923 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380571 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380583 4923 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380589 4923 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380595 4923 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380601 4923 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380606 4923 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380611 4923 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380617 4923 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380622 4923 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380627 4923 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380632 4923 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380637 4923 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380641 4923 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380646 4923 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380651 4923 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380657 4923 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380662 4923 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380666 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380671 4923 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380677 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380681 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380686 4923 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380691 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380696 4923 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380701 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380706 4923 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380715 4923 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380721 4923 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380727 4923 feature_gate.go:330] unrecognized feature gate: Example Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380734 4923 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380740 4923 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380746 4923 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380772 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380778 4923 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380782 4923 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380787 4923 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380794 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380800 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380807 4923 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380813 4923 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380819 4923 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380825 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380830 4923 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380835 4923 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380840 4923 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380845 4923 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380851 4923 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380857 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380864 4923 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380870 4923 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380877 4923 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380883 4923 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380891 4923 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380898 4923 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380904 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380910 4923 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380916 4923 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380922 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380928 4923 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380938 4923 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380946 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380953 4923 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380958 4923 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380964 4923 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380970 4923 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.380976 4923 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381914 4923 flags.go:64] FLAG: --address="0.0.0.0" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381933 4923 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381942 4923 flags.go:64] FLAG: --anonymous-auth="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381952 4923 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381959 4923 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381965 4923 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381976 4923 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381984 4923 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381990 4923 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.381996 4923 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382002 4923 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382008 4923 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382013 4923 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382019 4923 flags.go:64] FLAG: --cgroup-root="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382025 4923 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382033 4923 flags.go:64] FLAG: --client-ca-file="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382040 4923 flags.go:64] FLAG: --cloud-config="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382047 4923 flags.go:64] FLAG: --cloud-provider="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382054 4923 flags.go:64] FLAG: --cluster-dns="[]" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382064 4923 flags.go:64] FLAG: --cluster-domain="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382071 4923 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382078 4923 flags.go:64] FLAG: --config-dir="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382085 4923 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382093 4923 flags.go:64] FLAG: --container-log-max-files="5" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382105 4923 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382113 4923 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382124 4923 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382133 4923 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382142 4923 flags.go:64] FLAG: --contention-profiling="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382152 4923 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382159 4923 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382167 4923 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382175 4923 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382185 4923 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382193 4923 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382200 4923 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382206 4923 flags.go:64] FLAG: --enable-load-reader="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382211 4923 flags.go:64] FLAG: --enable-server="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382217 4923 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382224 4923 flags.go:64] FLAG: --event-burst="100" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382230 4923 flags.go:64] FLAG: --event-qps="50" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382235 4923 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382242 4923 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382248 4923 flags.go:64] FLAG: --eviction-hard="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382255 4923 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382261 4923 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382267 4923 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382272 4923 flags.go:64] FLAG: --eviction-soft="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382278 4923 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382283 4923 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382289 4923 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382295 4923 flags.go:64] FLAG: --experimental-mounter-path="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382300 4923 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382306 4923 flags.go:64] FLAG: --fail-swap-on="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382311 4923 flags.go:64] FLAG: --feature-gates="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382319 4923 flags.go:64] FLAG: --file-check-frequency="20s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382324 4923 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382330 4923 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382337 4923 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382343 4923 flags.go:64] FLAG: --healthz-port="10248" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382349 4923 flags.go:64] FLAG: --help="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382355 4923 flags.go:64] FLAG: --hostname-override="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382360 4923 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382366 4923 flags.go:64] FLAG: --http-check-frequency="20s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382372 4923 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382377 4923 flags.go:64] FLAG: --image-credential-provider-config="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382383 4923 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382388 4923 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382394 4923 flags.go:64] FLAG: --image-service-endpoint="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382399 4923 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382405 4923 flags.go:64] FLAG: --kube-api-burst="100" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382411 4923 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382416 4923 flags.go:64] FLAG: --kube-api-qps="50" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382422 4923 flags.go:64] FLAG: --kube-reserved="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382427 4923 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382433 4923 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382439 4923 flags.go:64] FLAG: --kubelet-cgroups="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382444 4923 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382450 4923 flags.go:64] FLAG: --lock-file="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382456 4923 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382461 4923 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382467 4923 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382475 4923 flags.go:64] FLAG: --log-json-split-stream="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382481 4923 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382487 4923 flags.go:64] FLAG: --log-text-split-stream="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382492 4923 flags.go:64] FLAG: --logging-format="text" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382498 4923 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382505 4923 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382511 4923 flags.go:64] FLAG: --manifest-url="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382516 4923 flags.go:64] FLAG: --manifest-url-header="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382525 4923 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382531 4923 flags.go:64] FLAG: --max-open-files="1000000" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382543 4923 flags.go:64] FLAG: --max-pods="110" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382549 4923 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382555 4923 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382561 4923 flags.go:64] FLAG: --memory-manager-policy="None" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382567 4923 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382573 4923 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382579 4923 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382584 4923 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382598 4923 flags.go:64] FLAG: --node-status-max-images="50" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382604 4923 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382610 4923 flags.go:64] FLAG: --oom-score-adj="-999" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382615 4923 flags.go:64] FLAG: --pod-cidr="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382620 4923 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382630 4923 flags.go:64] FLAG: --pod-manifest-path="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382636 4923 flags.go:64] FLAG: --pod-max-pids="-1" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382643 4923 flags.go:64] FLAG: --pods-per-core="0" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382649 4923 flags.go:64] FLAG: --port="10250" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382656 4923 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382662 4923 flags.go:64] FLAG: --provider-id="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382668 4923 flags.go:64] FLAG: --qos-reserved="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382674 4923 flags.go:64] FLAG: --read-only-port="10255" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382680 4923 flags.go:64] FLAG: --register-node="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382686 4923 flags.go:64] FLAG: --register-schedulable="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382692 4923 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382701 4923 flags.go:64] FLAG: --registry-burst="10" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382708 4923 flags.go:64] FLAG: --registry-qps="5" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382713 4923 flags.go:64] FLAG: --reserved-cpus="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382719 4923 flags.go:64] FLAG: --reserved-memory="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382726 4923 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382732 4923 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382738 4923 flags.go:64] FLAG: --rotate-certificates="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382744 4923 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382772 4923 flags.go:64] FLAG: --runonce="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382778 4923 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382784 4923 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382791 4923 flags.go:64] FLAG: --seccomp-default="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382797 4923 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382802 4923 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382808 4923 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382814 4923 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382819 4923 flags.go:64] FLAG: --storage-driver-password="root" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382825 4923 flags.go:64] FLAG: --storage-driver-secure="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382831 4923 flags.go:64] FLAG: --storage-driver-table="stats" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382836 4923 flags.go:64] FLAG: --storage-driver-user="root" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382842 4923 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382848 4923 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382854 4923 flags.go:64] FLAG: --system-cgroups="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382860 4923 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382868 4923 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382873 4923 flags.go:64] FLAG: --tls-cert-file="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382879 4923 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382886 4923 flags.go:64] FLAG: --tls-min-version="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382891 4923 flags.go:64] FLAG: --tls-private-key-file="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382897 4923 flags.go:64] FLAG: --topology-manager-policy="none" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382902 4923 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382907 4923 flags.go:64] FLAG: --topology-manager-scope="container" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382913 4923 flags.go:64] FLAG: --v="2" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382921 4923 flags.go:64] FLAG: --version="false" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382930 4923 flags.go:64] FLAG: --vmodule="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382937 4923 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.382943 4923 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383078 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383086 4923 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383091 4923 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383098 4923 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383105 4923 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383111 4923 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383117 4923 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383122 4923 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383128 4923 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383133 4923 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383138 4923 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383144 4923 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383148 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383153 4923 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383159 4923 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383163 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383170 4923 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383176 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383182 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383187 4923 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383192 4923 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383197 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383202 4923 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383208 4923 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383214 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383219 4923 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383225 4923 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383231 4923 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383238 4923 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383244 4923 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383251 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383256 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383262 4923 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383268 4923 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383277 4923 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383283 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383289 4923 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383294 4923 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383299 4923 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383306 4923 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383312 4923 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383317 4923 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383322 4923 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383327 4923 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383332 4923 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383337 4923 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383341 4923 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383347 4923 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383352 4923 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383357 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383362 4923 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383367 4923 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383372 4923 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383377 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383382 4923 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383386 4923 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383391 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383396 4923 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383401 4923 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383406 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383411 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383416 4923 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383421 4923 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383426 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383432 4923 feature_gate.go:330] unrecognized feature gate: Example Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383436 4923 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383442 4923 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383447 4923 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383452 4923 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383458 4923 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.383463 4923 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.383481 4923 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.395798 4923 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.395849 4923 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395936 4923 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395945 4923 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395949 4923 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395953 4923 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395957 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395962 4923 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395966 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395970 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395974 4923 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395978 4923 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395981 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395986 4923 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395992 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.395996 4923 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396000 4923 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396003 4923 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396007 4923 feature_gate.go:330] unrecognized feature gate: Example Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396010 4923 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396016 4923 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396021 4923 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396026 4923 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396030 4923 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396035 4923 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396049 4923 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396054 4923 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396058 4923 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396062 4923 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396066 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396069 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396073 4923 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396076 4923 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396081 4923 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396087 4923 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396091 4923 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396095 4923 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396100 4923 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396104 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396108 4923 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396112 4923 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396116 4923 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396119 4923 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396123 4923 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396127 4923 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396131 4923 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396136 4923 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396140 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396145 4923 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396149 4923 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396153 4923 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396156 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396160 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396164 4923 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396169 4923 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396173 4923 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396177 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396180 4923 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396184 4923 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396188 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396192 4923 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396195 4923 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396198 4923 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396202 4923 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396212 4923 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396216 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396219 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396222 4923 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396226 4923 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396229 4923 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396233 4923 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396237 4923 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396242 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.396250 4923 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396414 4923 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396427 4923 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396432 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396437 4923 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396442 4923 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396447 4923 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396454 4923 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396461 4923 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396468 4923 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396474 4923 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396480 4923 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396486 4923 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396490 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396495 4923 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396500 4923 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396505 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396510 4923 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396514 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396517 4923 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396521 4923 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396525 4923 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396529 4923 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396532 4923 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396536 4923 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396539 4923 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396543 4923 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396547 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396550 4923 feature_gate.go:330] unrecognized feature gate: Example Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396554 4923 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396558 4923 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396561 4923 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396566 4923 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396569 4923 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396573 4923 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396576 4923 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396580 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396584 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396588 4923 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396593 4923 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396598 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396602 4923 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396606 4923 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396609 4923 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396613 4923 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396617 4923 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396620 4923 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396624 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396628 4923 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396631 4923 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396635 4923 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396639 4923 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396643 4923 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396647 4923 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396650 4923 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396654 4923 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396658 4923 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396662 4923 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396666 4923 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396669 4923 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396673 4923 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396677 4923 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396681 4923 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396685 4923 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396689 4923 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396692 4923 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396697 4923 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396701 4923 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396706 4923 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396710 4923 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396714 4923 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.396718 4923 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.396725 4923 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.398685 4923 server.go:940] "Client rotation is on, will bootstrap in background" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.403863 4923 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.404003 4923 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.405537 4923 server.go:997] "Starting client certificate rotation" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.405563 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.406349 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-08 22:47:37.062197759 +0000 UTC Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.406454 4923 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 732h42m22.655746996s for next certificate rotation Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.430943 4923 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.435712 4923 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.452515 4923 log.go:25] "Validated CRI v1 runtime API" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.488305 4923 log.go:25] "Validated CRI v1 image API" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.490161 4923 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.495922 4923 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-09-09-38-12-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.495950 4923 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.512816 4923 manager.go:217] Machine: {Timestamp:2025-10-09 10:05:14.508654868 +0000 UTC m=+0.576836624 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:4c539cb4-9caf-4858-827b-2e25c2e1065c BootID:e7a319c8-81e7-4f84-9486-6dc3c791c3fc Filesystems:[{Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:2b:e4:d1 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:2b:e4:d1 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:3e:dc:b9 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9a:14:2f Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:60:c6:e2 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:2a:84:25 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:bb:e0:36 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:86:04:fa:93:a5:ce Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:7a:68:1c:c0:94:f9 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.513059 4923 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.513196 4923 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.515582 4923 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.515744 4923 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.515790 4923 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.515985 4923 topology_manager.go:138] "Creating topology manager with none policy" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.515997 4923 container_manager_linux.go:303] "Creating device plugin manager" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.516661 4923 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.517046 4923 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.518203 4923 state_mem.go:36] "Initialized new in-memory state store" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.518315 4923 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.522734 4923 kubelet.go:418] "Attempting to sync node with API server" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.522780 4923 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.522844 4923 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.522858 4923 kubelet.go:324] "Adding apiserver pod source" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.522931 4923 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.529263 4923 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.529934 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.530027 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.530105 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.530168 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.530491 4923 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.533452 4923 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535214 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535301 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535386 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535438 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535507 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535578 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535628 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535680 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535728 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.535942 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.536041 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.536114 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.538731 4923 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.539266 4923 server.go:1280] "Started kubelet" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541130 4923 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541132 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541178 4923 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 09:32:23.165274551 +0000 UTC Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541209 4923 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1775h27m8.624068031s for next certificate rotation Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541362 4923 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541361 4923 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541445 4923 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541459 4923 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.541470 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541359 4923 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541742 4923 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.541871 4923 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 09 10:05:14 crc systemd[1]: Started Kubernetes Kubelet. Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.542287 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.542354 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.542834 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.120:6443: connect: connection refused" interval="200ms" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.543271 4923 server.go:460] "Adding debug handlers to kubelet server" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.546412 4923 factory.go:55] Registering systemd factory Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.546446 4923 factory.go:221] Registration of the systemd container factory successfully Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.548356 4923 factory.go:153] Registering CRI-O factory Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.548396 4923 factory.go:221] Registration of the crio container factory successfully Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.548468 4923 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.548490 4923 factory.go:103] Registering Raw factory Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.548505 4923 manager.go:1196] Started watching for new ooms in manager Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.549439 4923 manager.go:319] Starting recovery of all containers Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.548812 4923 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.120:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186cca9963441f29 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-09 10:05:14.539237161 +0000 UTC m=+0.607418917,LastTimestamp:2025-10-09 10:05:14.539237161 +0000 UTC m=+0.607418917,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563450 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563512 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563533 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563548 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563560 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563579 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563590 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563601 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563613 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563622 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563633 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563643 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563653 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563665 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563674 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563685 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563694 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563710 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563719 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563728 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563737 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563779 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563794 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563804 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563816 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563828 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563839 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563849 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563859 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563886 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563897 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563908 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563932 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563941 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563952 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563961 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563987 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.563998 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564009 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564019 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564105 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564115 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564125 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564135 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564146 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564157 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564168 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564180 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564208 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564218 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564227 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564237 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564250 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564260 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564271 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564280 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564294 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564308 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564316 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564331 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564340 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564348 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564357 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564366 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564374 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564382 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564394 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564403 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564411 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564420 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564429 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564444 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564455 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564463 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564479 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564491 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564505 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564516 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564527 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564539 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564551 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564562 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564573 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564593 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564604 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564614 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564624 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564636 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564650 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564662 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564675 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564690 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564700 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.564710 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569105 4923 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569156 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569174 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569185 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569199 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569210 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569221 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569236 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569250 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569267 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569285 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569312 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569328 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569345 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569358 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569371 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569382 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569415 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569428 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569441 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569454 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569466 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569482 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569496 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569506 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569516 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569526 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569537 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569547 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569557 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569568 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569578 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569589 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569602 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569612 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569624 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569635 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569646 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569656 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569666 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569708 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569720 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569735 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569746 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569772 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569782 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569797 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569808 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569823 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569836 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569846 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569856 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569866 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569878 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569892 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569901 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569911 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569920 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569930 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569944 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569955 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569968 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569978 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569987 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.569998 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570009 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570020 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570031 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570041 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570051 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570061 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570078 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570088 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570099 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570108 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570117 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570127 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570147 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570157 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570168 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570179 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570190 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570201 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570211 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570221 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570231 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570241 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570252 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570266 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570277 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570288 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570299 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570309 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570327 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570337 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570347 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570357 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570368 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570378 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570388 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570404 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570418 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570429 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570439 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570451 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570461 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570472 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570482 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570494 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570504 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570514 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570524 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570536 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570547 4923 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570556 4923 reconstruct.go:97] "Volume reconstruction finished" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.570564 4923 reconciler.go:26] "Reconciler: start to sync state" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.577990 4923 manager.go:324] Recovery completed Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.587874 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.590075 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.590107 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.590117 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.590725 4923 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.590744 4923 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.590789 4923 state_mem.go:36] "Initialized new in-memory state store" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.598077 4923 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.599968 4923 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.600012 4923 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.600049 4923 kubelet.go:2335] "Starting kubelet main sync loop" Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.600096 4923 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 09 10:05:14 crc kubenswrapper[4923]: W1009 10:05:14.603415 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.604102 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.620080 4923 policy_none.go:49] "None policy: Start" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.621392 4923 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.621420 4923 state_mem.go:35] "Initializing new in-memory state store" Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.642321 4923 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.671216 4923 manager.go:334] "Starting Device Plugin manager" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.671370 4923 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.671381 4923 server.go:79] "Starting device plugin registration server" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.671920 4923 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.671936 4923 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.672205 4923 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.672326 4923 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.672336 4923 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.680027 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.700478 4923 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.700650 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.701973 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.702010 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.702021 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.702185 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.702957 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.703035 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.705526 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.705559 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.705570 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.705527 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.705642 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.705656 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.705682 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.705868 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.705919 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.706846 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.706880 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.706892 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.707787 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.707817 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.707829 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.707937 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708201 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708231 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708573 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708601 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708612 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708706 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708818 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708840 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708929 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708957 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.708965 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.709429 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.709461 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.709474 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.709653 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.709690 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.710040 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.710068 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.710076 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.710652 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.710735 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.710870 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.743575 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.120:6443: connect: connection refused" interval="400ms" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.772904 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.772967 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.773710 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.773734 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.773769 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.773785 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.773839 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.773858 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.773942 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.774002 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.774026 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.774061 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.774086 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.774108 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.774130 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.774144 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.775082 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.775108 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.775117 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.775140 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.775468 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.120:6443: connect: connection refused" node="crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875617 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875693 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875714 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875736 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875784 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875806 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875824 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875860 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875877 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875868 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875897 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875963 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875970 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875974 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875996 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876018 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.875895 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876039 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876101 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876133 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876141 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876165 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876195 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876207 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876211 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876228 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876223 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876195 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876262 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.876337 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.975931 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.977086 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.977125 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.977135 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:14 crc kubenswrapper[4923]: I1009 10:05:14.977156 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 10:05:14 crc kubenswrapper[4923]: E1009 10:05:14.977542 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.120:6443: connect: connection refused" node="crc" Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.041594 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.057075 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.074063 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 10:05:15 crc kubenswrapper[4923]: W1009 10:05:15.090970 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c6fb7e90a0cc0af19a237235d0b56dc510be70a9f9caaf328cce16e74ac3d563 WatchSource:0}: Error finding container c6fb7e90a0cc0af19a237235d0b56dc510be70a9f9caaf328cce16e74ac3d563: Status 404 returned error can't find the container with id c6fb7e90a0cc0af19a237235d0b56dc510be70a9f9caaf328cce16e74ac3d563 Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.092098 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 09 10:05:15 crc kubenswrapper[4923]: W1009 10:05:15.094453 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-efa6261c353c7a37533a22f0519dc1f8983cbd3cad17379d7cf1b145f6d468af WatchSource:0}: Error finding container efa6261c353c7a37533a22f0519dc1f8983cbd3cad17379d7cf1b145f6d468af: Status 404 returned error can't find the container with id efa6261c353c7a37533a22f0519dc1f8983cbd3cad17379d7cf1b145f6d468af Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.098345 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:15 crc kubenswrapper[4923]: W1009 10:05:15.104081 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-82387affb510cdf24de31f0c6f0bc64159c296d723a7cf8ec7bc2e398b177cd8 WatchSource:0}: Error finding container 82387affb510cdf24de31f0c6f0bc64159c296d723a7cf8ec7bc2e398b177cd8: Status 404 returned error can't find the container with id 82387affb510cdf24de31f0c6f0bc64159c296d723a7cf8ec7bc2e398b177cd8 Oct 09 10:05:15 crc kubenswrapper[4923]: E1009 10:05:15.144576 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.120:6443: connect: connection refused" interval="800ms" Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.377832 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.380355 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.380398 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.380407 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.380434 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 10:05:15 crc kubenswrapper[4923]: E1009 10:05:15.380988 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.120:6443: connect: connection refused" node="crc" Oct 09 10:05:15 crc kubenswrapper[4923]: W1009 10:05:15.480808 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:15 crc kubenswrapper[4923]: E1009 10:05:15.481081 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.542420 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.604996 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c3eec2f60c16a3462f724ac32bf7e481f27ecf1cfc5a7a9085a680437568fac3"} Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.605974 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3db9a8a454ab06cbd0f5e8457e8c9e15faec42d5b09ac39954a948a5cbf9f2c5"} Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.606915 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"82387affb510cdf24de31f0c6f0bc64159c296d723a7cf8ec7bc2e398b177cd8"} Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.608179 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c6fb7e90a0cc0af19a237235d0b56dc510be70a9f9caaf328cce16e74ac3d563"} Oct 09 10:05:15 crc kubenswrapper[4923]: I1009 10:05:15.609176 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"efa6261c353c7a37533a22f0519dc1f8983cbd3cad17379d7cf1b145f6d468af"} Oct 09 10:05:15 crc kubenswrapper[4923]: W1009 10:05:15.871226 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:15 crc kubenswrapper[4923]: E1009 10:05:15.871332 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:15 crc kubenswrapper[4923]: W1009 10:05:15.924124 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:15 crc kubenswrapper[4923]: E1009 10:05:15.924214 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:15 crc kubenswrapper[4923]: E1009 10:05:15.946423 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.120:6443: connect: connection refused" interval="1.6s" Oct 09 10:05:16 crc kubenswrapper[4923]: W1009 10:05:16.122155 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:16 crc kubenswrapper[4923]: E1009 10:05:16.122259 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.181829 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.183651 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.183704 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.183717 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.183741 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 10:05:16 crc kubenswrapper[4923]: E1009 10:05:16.184408 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.120:6443: connect: connection refused" node="crc" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.542691 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.616372 4923 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="67fc80c6b70b9e3c1ff2221ad07f802023f327b9d7a0ca381d6ab978ea9a2212" exitCode=0 Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.616519 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"67fc80c6b70b9e3c1ff2221ad07f802023f327b9d7a0ca381d6ab978ea9a2212"} Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.616592 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.619026 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.619069 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.619082 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.620343 4923 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="8ad87a876197be3010d30a811c8e4f48c94109e6125a78e4a0b13316689c6058" exitCode=0 Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.620419 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"8ad87a876197be3010d30a811c8e4f48c94109e6125a78e4a0b13316689c6058"} Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.620431 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.621304 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.621329 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.621337 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.623887 4923 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572" exitCode=0 Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.623944 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572"} Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.623956 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.625056 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.625084 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.625093 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.626908 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410"} Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.627011 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f"} Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.627095 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d"} Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.627117 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa"} Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.627294 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.628506 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.628541 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.628551 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.629176 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133" exitCode=0 Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.629217 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133"} Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.629238 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.629984 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.630006 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.630015 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.631138 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.631796 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.631822 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:16 crc kubenswrapper[4923]: I1009 10:05:16.631832 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.541780 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:17 crc kubenswrapper[4923]: E1009 10:05:17.548281 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.120:6443: connect: connection refused" interval="3.2s" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.641928 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205"} Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.641972 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655"} Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.641985 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2"} Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.643980 4923 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="4317af6a6232cbb629622b3a324a21e4452281d81665c2086d9cccd6e60eace1" exitCode=0 Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.644043 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"4317af6a6232cbb629622b3a324a21e4452281d81665c2086d9cccd6e60eace1"} Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.644209 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.645571 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"cb2144f20e96788c6d2ea0fa749721bedc39b29cbd34dbef062f8d4517805d8b"} Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.645635 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.645674 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.645706 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.645717 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.646508 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.646529 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.646539 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.649089 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"2e9c6706f9f71acc1376c0a0def4c117db567dc5b60962106e6d0aebb43caaa6"} Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.649116 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ae77d7fd42982b8c1531d352ed15257c3626fe61672015a7ea3a1c7b88db9ae0"} Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.649135 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8eef9abe2735aef05ea4fb2ea269f806a4b4e4cc75d890aedc26b61e00890380"} Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.649155 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.649157 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.650007 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.650042 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.650056 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.650103 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.650129 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.650141 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.784693 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.786871 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.786931 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.786949 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:17 crc kubenswrapper[4923]: I1009 10:05:17.786983 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 10:05:17 crc kubenswrapper[4923]: E1009 10:05:17.787632 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.120:6443: connect: connection refused" node="crc" Oct 09 10:05:17 crc kubenswrapper[4923]: W1009 10:05:17.800310 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:17 crc kubenswrapper[4923]: E1009 10:05:17.800414 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:18 crc kubenswrapper[4923]: W1009 10:05:18.091694 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:18 crc kubenswrapper[4923]: E1009 10:05:18.091850 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:18 crc kubenswrapper[4923]: E1009 10:05:18.342525 4923 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.120:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186cca9963441f29 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-09 10:05:14.539237161 +0000 UTC m=+0.607418917,LastTimestamp:2025-10-09 10:05:14.539237161 +0000 UTC m=+0.607418917,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.542968 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.656747 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4"} Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.656865 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74"} Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.656799 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.658175 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.658214 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.658223 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.660732 4923 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c5ca0b5b151a811bc92b070a05fe4715806b3aaac5b126edcf3b857826f1b0a8" exitCode=0 Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.660786 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c5ca0b5b151a811bc92b070a05fe4715806b3aaac5b126edcf3b857826f1b0a8"} Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.660843 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.660865 4923 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.660896 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.660915 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.662057 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.662089 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.662057 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.662109 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.662098 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.662119 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.662190 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.662230 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.662196 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:18 crc kubenswrapper[4923]: W1009 10:05:18.761260 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:18 crc kubenswrapper[4923]: E1009 10:05:18.761369 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.797151 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.797329 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.798315 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.798339 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.798347 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:18 crc kubenswrapper[4923]: I1009 10:05:18.803349 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:18 crc kubenswrapper[4923]: W1009 10:05:18.871289 4923 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.120:6443: connect: connection refused Oct 09 10:05:18 crc kubenswrapper[4923]: E1009 10:05:18.871390 4923 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.120:6443: connect: connection refused" logger="UnhandledError" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.092070 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.132931 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.665201 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.665843 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3c3c69ab95d9e0c9169dc181f8d3d0c47e714f83c03c0ab8285f1541a38dfc55"} Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.665886 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.665897 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a9c609494207d75780343901ccab8d445d3edc47e8caed456a9888fbd200610e"} Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.665906 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ef598b437242c992a46d51a181ceb26e2148d31c34b55eee232f835e4d044735"} Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.665914 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"35fd37b9e6ff82e1c753c1ea4aee0428eb0fb4f16d14d5d7bcd02f47f1d54750"} Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.665970 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.666318 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.666942 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.666971 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.666981 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.667592 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.667612 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.667630 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.668008 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.668025 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:19 crc kubenswrapper[4923]: I1009 10:05:19.668033 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.675283 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"03af1ab0936b4c97aee4f5b00fca091dd6685eecf8ec15a58098eea4cb26e52b"} Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.675387 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.675392 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.677652 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.677696 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.677709 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.678435 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.678469 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.678479 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.988553 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.993890 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.994059 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.994085 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:20 crc kubenswrapper[4923]: I1009 10:05:20.994145 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 10:05:21 crc kubenswrapper[4923]: I1009 10:05:21.678664 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:21 crc kubenswrapper[4923]: I1009 10:05:21.680247 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:21 crc kubenswrapper[4923]: I1009 10:05:21.680295 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:21 crc kubenswrapper[4923]: I1009 10:05:21.680312 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:22 crc kubenswrapper[4923]: I1009 10:05:22.136984 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:22 crc kubenswrapper[4923]: I1009 10:05:22.137289 4923 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 10:05:22 crc kubenswrapper[4923]: I1009 10:05:22.137380 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:22 crc kubenswrapper[4923]: I1009 10:05:22.139506 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:22 crc kubenswrapper[4923]: I1009 10:05:22.139558 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:22 crc kubenswrapper[4923]: I1009 10:05:22.139578 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.230643 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.232122 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.234289 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.234387 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.234408 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.385305 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.385547 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.387008 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.387062 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.387071 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.996589 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.996815 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.997893 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.997938 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:23 crc kubenswrapper[4923]: I1009 10:05:23.997948 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:24 crc kubenswrapper[4923]: I1009 10:05:24.165848 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:24 crc kubenswrapper[4923]: I1009 10:05:24.166140 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:24 crc kubenswrapper[4923]: I1009 10:05:24.167858 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:24 crc kubenswrapper[4923]: I1009 10:05:24.167914 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:24 crc kubenswrapper[4923]: I1009 10:05:24.167924 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:24 crc kubenswrapper[4923]: E1009 10:05:24.680396 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.137000 4923 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.137097 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.342981 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.343273 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.345375 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.345533 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.345699 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.351108 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.689084 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.690011 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.690051 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:25 crc kubenswrapper[4923]: I1009 10:05:25.690061 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:29 crc kubenswrapper[4923]: I1009 10:05:29.543204 4923 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Oct 09 10:05:29 crc kubenswrapper[4923]: I1009 10:05:29.879944 4923 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 09 10:05:29 crc kubenswrapper[4923]: I1009 10:05:29.880029 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 09 10:05:29 crc kubenswrapper[4923]: I1009 10:05:29.885363 4923 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Oct 09 10:05:29 crc kubenswrapper[4923]: I1009 10:05:29.885487 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.391359 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.391515 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.392707 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.392779 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.392797 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.395326 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.712394 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.714465 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.714517 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:33 crc kubenswrapper[4923]: I1009 10:05:33.714532 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.028863 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.029039 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.030242 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.030283 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.030294 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.048432 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 09 10:05:34 crc kubenswrapper[4923]: E1009 10:05:34.680640 4923 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.715123 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.716564 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.716671 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.716700 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:34 crc kubenswrapper[4923]: E1009 10:05:34.880142 4923 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.884739 4923 trace.go:236] Trace[124269700]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Oct-2025 10:05:21.656) (total time: 13228ms): Oct 09 10:05:34 crc kubenswrapper[4923]: Trace[124269700]: ---"Objects listed" error: 13228ms (10:05:34.884) Oct 09 10:05:34 crc kubenswrapper[4923]: Trace[124269700]: [13.228514005s] [13.228514005s] END Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.884824 4923 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.885061 4923 trace.go:236] Trace[1315982678]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Oct-2025 10:05:23.302) (total time: 11582ms): Oct 09 10:05:34 crc kubenswrapper[4923]: Trace[1315982678]: ---"Objects listed" error: 11582ms (10:05:34.884) Oct 09 10:05:34 crc kubenswrapper[4923]: Trace[1315982678]: [11.582337552s] [11.582337552s] END Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.885096 4923 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.885390 4923 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.885479 4923 trace.go:236] Trace[1870566638]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Oct-2025 10:05:23.514) (total time: 11370ms): Oct 09 10:05:34 crc kubenswrapper[4923]: Trace[1870566638]: ---"Objects listed" error: 11370ms (10:05:34.885) Oct 09 10:05:34 crc kubenswrapper[4923]: Trace[1870566638]: [11.370954819s] [11.370954819s] END Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.885511 4923 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.885712 4923 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 09 10:05:34 crc kubenswrapper[4923]: E1009 10:05:34.893885 4923 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.936815 4923 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38118->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.936874 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:38118->192.168.126.11:17697: read: connection reset by peer" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.937225 4923 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.937299 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.960865 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:34 crc kubenswrapper[4923]: I1009 10:05:34.976635 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.239016 4923 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.239081 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.536692 4923 apiserver.go:52] "Watching apiserver" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.539404 4923 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.539721 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.540396 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.540491 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.540685 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.541007 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.541038 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.541081 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.541106 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.541148 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.541387 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.542964 4923 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.543359 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.543852 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.543896 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.543945 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.544080 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.544410 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.544422 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.545311 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.546880 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.576846 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.590692 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.590784 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.590879 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.590905 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.590957 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.590979 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591079 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591225 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591275 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591287 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591544 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591904 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591931 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591936 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591320 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.591991 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592001 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592025 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592048 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592070 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592262 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592411 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592421 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592442 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592465 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592500 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592523 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592567 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592725 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592932 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592988 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593103 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.592592 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593174 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593197 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593391 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593437 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593455 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593660 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593716 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593738 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593843 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.593972 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594188 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594249 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594258 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594271 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594287 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594306 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594350 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594366 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594384 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594403 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594421 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594439 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594483 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594806 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594839 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.594963 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595089 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595162 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595234 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595258 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595338 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595304 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595366 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595618 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595865 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595916 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595941 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595973 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.595993 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596017 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596038 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596057 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596075 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596097 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596120 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596144 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596176 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596205 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596242 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596299 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596321 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596341 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596357 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596373 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596389 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596412 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596429 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596450 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596472 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596487 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596504 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596553 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596577 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596596 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596626 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596648 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596670 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596690 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596714 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596731 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596762 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596787 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596808 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596825 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596840 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596855 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596870 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596886 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596901 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596915 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596930 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596945 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596960 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596975 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596991 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596340 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.603705 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.599613 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.603964 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596626 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596911 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596960 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.598253 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.599299 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.599537 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.599965 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.604406 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.596582 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.600191 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.600100 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.600243 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.600882 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.599623 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.600063 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.601267 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.601593 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.601437 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.601661 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.601609 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.602040 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.602342 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.602472 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.602505 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.602497 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605409 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.602919 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.603031 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.603040 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.603127 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.603378 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.603411 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.603434 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.604906 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.604022 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.604987 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605474 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605043 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605011 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605619 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605662 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605697 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605730 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605746 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605871 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605944 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.606012 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.606045 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.606342 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.606492 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.606664 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.606831 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.605787 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.601031 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607006 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607013 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607102 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607129 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607150 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607444 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607622 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607684 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607692 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.607945 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.608938 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.608973 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.608999 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.608985 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609023 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609055 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609078 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609103 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609126 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609144 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609239 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609260 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609358 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609373 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609540 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609590 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609627 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.609973 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.610076 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.610640 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.610694 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.610715 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.611598 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:05:36.11157265 +0000 UTC m=+22.179754406 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.611844 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.611889 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.612006 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.612132 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.612815 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.612831 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.612856 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.612886 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613316 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613362 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613391 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613414 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613440 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613465 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613489 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613512 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613536 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613560 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613584 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613609 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613634 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613656 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613682 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613705 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613683 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613728 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613771 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613797 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613823 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613848 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613873 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613897 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613921 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613948 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.613976 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614000 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614022 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614046 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614121 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614225 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614256 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614281 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614307 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614332 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614329 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614358 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614386 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614414 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614440 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614466 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.612366 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614492 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614556 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614589 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614617 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614643 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614668 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614694 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614720 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614726 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614743 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614790 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614816 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614903 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614963 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.614993 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615019 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615046 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615073 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615097 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615124 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615178 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615207 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615236 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615267 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615273 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615302 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615331 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615340 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615356 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615384 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615412 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615417 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615439 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615471 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615498 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615554 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615583 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615590 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615610 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615634 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615722 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615734 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.615978 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616068 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616211 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616276 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616294 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616339 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616357 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616374 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616390 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616439 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616459 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616469 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616478 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616497 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616516 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616540 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616561 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616636 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616641 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616666 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616670 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616743 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616763 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616798 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616825 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616845 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616865 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.616907 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.617043 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.617332 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.617331 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.617616 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.617719 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.617798 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.618069 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.618251 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.618436 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.618467 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.618617 4923 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.619447 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.619887 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.620475 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.621047 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.621514 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.621544 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.621625 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.622092 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.622118 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.622535 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.622958 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.622855 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.623393 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.623689 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.623881 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.624182 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.624443 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.624712 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.624737 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.625030 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.625092 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.625216 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.625619 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.625959 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.625988 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.627087 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.627642 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.627996 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.629458 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.629941 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.630403 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.631498 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.631863 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.632114 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.631996 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.632393 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.632679 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.633016 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.633350 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.632553 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.633654 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.633708 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.633893 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.633969 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.634039 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.634059 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.634061 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.634074 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.634683 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.634742 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.634786 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.634791 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:36.134732061 +0000 UTC m=+22.202913827 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.634924 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:36.134902875 +0000 UTC m=+22.203084621 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635012 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635246 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635331 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635332 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.635400 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:36.135391729 +0000 UTC m=+22.203573485 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635481 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635502 4923 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635515 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635528 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635559 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635573 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635585 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635631 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635639 4923 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635680 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635716 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635787 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.635810 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636120 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636129 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636134 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636236 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636248 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636258 4923 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636269 4923 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636278 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636288 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636302 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636318 4923 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636332 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636349 4923 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636364 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636375 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636385 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636394 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636404 4923 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636413 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636428 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636437 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636447 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636457 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636465 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636477 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636486 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636620 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636703 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636716 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636726 4923 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636738 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.636978 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637094 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637204 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637190 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637442 4923 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637461 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637485 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637616 4923 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637630 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637640 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637652 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637663 4923 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637672 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637681 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637691 4923 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637699 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637710 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637722 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637733 4923 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637747 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637775 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637786 4923 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637800 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637811 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637821 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637835 4923 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637524 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637848 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637590 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637886 4923 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637900 4923 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637914 4923 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637949 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637959 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637968 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637977 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.637987 4923 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638021 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638047 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638057 4923 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638066 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638077 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638109 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638118 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638126 4923 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638135 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638144 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638153 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638186 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638195 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638203 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638212 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638223 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638231 4923 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638258 4923 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638268 4923 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638276 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638278 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638287 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638313 4923 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638325 4923 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638336 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638347 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638357 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638366 4923 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638375 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638387 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638399 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638415 4923 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638427 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638439 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638450 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638461 4923 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638474 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638485 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638497 4923 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638508 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638519 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638529 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638538 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638554 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638562 4923 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638572 4923 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638582 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638592 4923 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.638603 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.639615 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.640089 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.643213 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.645501 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.645680 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.645747 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.646397 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.646426 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.646442 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.646510 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:36.146485565 +0000 UTC m=+22.214667511 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.648075 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.648534 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.653898 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.654077 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.659137 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.665502 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.665847 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.668594 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.672339 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.675036 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.682284 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.696604 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.708493 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.719334 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.724169 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.727008 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4"} Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.727961 4923 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4" exitCode=255 Oct 09 10:05:35 crc kubenswrapper[4923]: E1009 10:05:35.734494 4923 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.734490 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739125 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739191 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739236 4923 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739254 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739273 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739288 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739301 4923 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739314 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739327 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739341 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739353 4923 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739380 4923 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739396 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739407 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739420 4923 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739428 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739437 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739435 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739448 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739396 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739548 4923 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739581 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739607 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739619 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739631 4923 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739645 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739660 4923 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739673 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739685 4923 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739696 4923 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739708 4923 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739723 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739767 4923 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739801 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739814 4923 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739829 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739843 4923 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739855 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739866 4923 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739877 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739889 4923 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.739900 4923 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740119 4923 scope.go:117] "RemoveContainer" containerID="0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740282 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740297 4923 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740338 4923 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740351 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740363 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740373 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740383 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740394 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740407 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740417 4923 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740426 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740434 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740442 4923 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740453 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740461 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740470 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740479 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740487 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740495 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740506 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740516 4923 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740527 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740541 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740554 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740563 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740572 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740585 4923 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740603 4923 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740614 4923 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740626 4923 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740637 4923 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740648 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740659 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740671 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740685 4923 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740695 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740704 4923 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740723 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.740732 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.748144 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.770893 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.785061 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.798789 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.813234 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.835141 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.849824 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.852880 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.860584 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.868650 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.875827 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.906917 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:35 crc kubenswrapper[4923]: I1009 10:05:35.936046 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.143267 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.143380 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:05:37.143359776 +0000 UTC m=+23.211541532 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.143511 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.143623 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.143672 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:37.143663825 +0000 UTC m=+23.211845581 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.143689 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.143726 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.143732 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.143794 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:37.143784848 +0000 UTC m=+23.211966594 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.143846 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.143859 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.143871 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.143905 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:37.143897611 +0000 UTC m=+23.212079367 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.245197 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.245481 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.245553 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.245569 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:36 crc kubenswrapper[4923]: E1009 10:05:36.245651 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:37.245630924 +0000 UTC m=+23.313812670 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.363079 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-nrck8"] Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.363409 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nrck8" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.366691 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.366930 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.367015 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.382593 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.393611 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.406307 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.424425 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.447045 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.470663 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.504161 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.518014 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.530649 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.548019 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7fd8165c-555f-4ded-8081-cb273d195245-hosts-file\") pod \"node-resolver-nrck8\" (UID: \"7fd8165c-555f-4ded-8081-cb273d195245\") " pod="openshift-dns/node-resolver-nrck8" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.548090 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs4hr\" (UniqueName: \"kubernetes.io/projected/7fd8165c-555f-4ded-8081-cb273d195245-kube-api-access-xs4hr\") pod \"node-resolver-nrck8\" (UID: \"7fd8165c-555f-4ded-8081-cb273d195245\") " pod="openshift-dns/node-resolver-nrck8" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.604354 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.604997 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.606676 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.607483 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.608654 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.609226 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.610015 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.611199 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.611991 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.613217 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.613900 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.615279 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.615922 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.616613 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.618343 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.618896 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.619696 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.620215 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.620949 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.621553 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.622062 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.622602 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.623051 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.623797 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.624210 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.624876 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.625531 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.626042 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.627611 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.628251 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.628886 4923 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.629064 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.630771 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.631390 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.632148 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.636089 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.637345 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.638043 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.639419 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.640294 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.641436 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.642196 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.643432 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.644233 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.645262 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.645940 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.647038 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.648035 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.648451 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7fd8165c-555f-4ded-8081-cb273d195245-hosts-file\") pod \"node-resolver-nrck8\" (UID: \"7fd8165c-555f-4ded-8081-cb273d195245\") " pod="openshift-dns/node-resolver-nrck8" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.648515 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs4hr\" (UniqueName: \"kubernetes.io/projected/7fd8165c-555f-4ded-8081-cb273d195245-kube-api-access-xs4hr\") pod \"node-resolver-nrck8\" (UID: \"7fd8165c-555f-4ded-8081-cb273d195245\") " pod="openshift-dns/node-resolver-nrck8" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.648674 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/7fd8165c-555f-4ded-8081-cb273d195245-hosts-file\") pod \"node-resolver-nrck8\" (UID: \"7fd8165c-555f-4ded-8081-cb273d195245\") " pod="openshift-dns/node-resolver-nrck8" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.649087 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.649666 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.650731 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.651422 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.652171 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.653203 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.677860 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs4hr\" (UniqueName: \"kubernetes.io/projected/7fd8165c-555f-4ded-8081-cb273d195245-kube-api-access-xs4hr\") pod \"node-resolver-nrck8\" (UID: \"7fd8165c-555f-4ded-8081-cb273d195245\") " pod="openshift-dns/node-resolver-nrck8" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.732817 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.735204 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2"} Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.735556 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.736137 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6e8f6413e76b2d97f9c86e9f725f8289288d0c71c4060932c9e022aad768b013"} Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.737664 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e"} Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.737698 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da"} Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.737712 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8894ee6c005cb13f2a26e427aef3137a9e5009e2636aa06087fec7416144eb74"} Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.738963 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c"} Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.739010 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b428aefeef1d44f10d18ce8d9d81a3443a4c4b97ae020bd25059b45d22ed409b"} Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.743864 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-75qcv"] Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.744773 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.746714 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-j9tk4"] Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.747229 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.749359 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.749611 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.749770 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.749865 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.749902 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.749882 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.749963 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.750204 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.750215 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.750811 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.750917 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.753363 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-4cn4f"] Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.753817 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.753955 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-frh4j"] Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.754161 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.754218 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.755628 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.756979 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.757159 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.757229 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.757416 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.757614 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.757634 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.757671 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.770210 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.790395 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.800516 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.810419 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.820942 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.849943 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-os-release\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850005 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-node-log\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850026 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-bin\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850122 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-log-socket\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850160 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-etc-openvswitch\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850180 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-config\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850203 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-systemd\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850226 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-kubelet\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850245 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-openvswitch\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850263 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-ovn\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850282 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc2a175c-508c-4f50-8205-013a9482c989-ovn-node-metrics-cert\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850304 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-cnibin\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850336 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-netns\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850356 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-systemd-units\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850377 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-ovn-kubernetes\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850397 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-netd\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850418 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850437 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-env-overrides\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850458 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-system-cni-dir\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850480 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v9j6\" (UniqueName: \"kubernetes.io/projected/bc2a175c-508c-4f50-8205-013a9482c989-kube-api-access-8v9j6\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850533 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-var-lib-openvswitch\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850557 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/99b8492d-66ff-496d-8514-5d50a498e04a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850592 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/99b8492d-66ff-496d-8514-5d50a498e04a-cni-binary-copy\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850615 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-987lm\" (UniqueName: \"kubernetes.io/projected/99b8492d-66ff-496d-8514-5d50a498e04a-kube-api-access-987lm\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850644 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-slash\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850664 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-script-lib\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.850699 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.857095 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.883132 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:36Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.900691 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:36Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.918742 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:36Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.929154 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:36Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.940100 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:36Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951551 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-netns\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951650 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1130e920-d4a8-44fe-8bb4-050213d1cff2-mcd-auth-proxy-config\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951703 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-run-k8s-cni-cncf-io\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951739 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-conf-dir\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951777 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-systemd-units\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951794 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-ovn-kubernetes\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951818 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-netd\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951834 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951849 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-env-overrides\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951865 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-system-cni-dir\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951881 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v9j6\" (UniqueName: \"kubernetes.io/projected/bc2a175c-508c-4f50-8205-013a9482c989-kube-api-access-8v9j6\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951898 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1130e920-d4a8-44fe-8bb4-050213d1cff2-proxy-tls\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951914 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-daemon-config\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951944 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-var-lib-openvswitch\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951968 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-socket-dir-parent\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.951989 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-hostroot\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952009 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-etc-kubernetes\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952034 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/99b8492d-66ff-496d-8514-5d50a498e04a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952061 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/99b8492d-66ff-496d-8514-5d50a498e04a-cni-binary-copy\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952085 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-987lm\" (UniqueName: \"kubernetes.io/projected/99b8492d-66ff-496d-8514-5d50a498e04a-kube-api-access-987lm\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952119 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-slash\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952141 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-script-lib\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952178 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952202 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-run-netns\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952225 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-os-release\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952250 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-node-log\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952272 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-bin\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952294 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/1130e920-d4a8-44fe-8bb4-050213d1cff2-rootfs\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952318 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-var-lib-cni-bin\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952341 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-var-lib-cni-multus\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952364 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-var-lib-kubelet\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952388 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-run-multus-certs\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952431 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-log-socket\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952455 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzq58\" (UniqueName: \"kubernetes.io/projected/1130e920-d4a8-44fe-8bb4-050213d1cff2-kube-api-access-nzq58\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952479 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-etc-openvswitch\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952503 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-config\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952531 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-systemd\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952553 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-system-cni-dir\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952576 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-cnibin\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952601 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khbl2\" (UniqueName: \"kubernetes.io/projected/d5c5f455-4e94-4f9a-b6fa-2535685c6354-kube-api-access-khbl2\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952601 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-ovn-kubernetes\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952624 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-kubelet\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952670 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-kubelet\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952676 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-cni-dir\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952693 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-os-release\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952718 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-openvswitch\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952733 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-ovn\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952771 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc2a175c-508c-4f50-8205-013a9482c989-ovn-node-metrics-cert\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952787 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d5c5f455-4e94-4f9a-b6fa-2535685c6354-cni-binary-copy\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952805 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-cnibin\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952844 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-cnibin\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952843 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-netns\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952871 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-netd\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952898 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-openvswitch\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952910 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952919 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-ovn\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.952568 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-systemd-units\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.953364 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-env-overrides\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.953406 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-system-cni-dir\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.953401 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/99b8492d-66ff-496d-8514-5d50a498e04a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.953430 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-slash\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.953849 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-script-lib\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.953924 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-var-lib-openvswitch\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.954129 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-os-release\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.954175 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-node-log\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.954202 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-bin\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.954206 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/99b8492d-66ff-496d-8514-5d50a498e04a-cni-binary-copy\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.954250 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-log-socket\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.954278 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-etc-openvswitch\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.954312 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-systemd\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.954318 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/99b8492d-66ff-496d-8514-5d50a498e04a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.954443 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-config\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.957623 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc2a175c-508c-4f50-8205-013a9482c989-ovn-node-metrics-cert\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.959296 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:36Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.968879 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v9j6\" (UniqueName: \"kubernetes.io/projected/bc2a175c-508c-4f50-8205-013a9482c989-kube-api-access-8v9j6\") pod \"ovnkube-node-75qcv\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.971654 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-987lm\" (UniqueName: \"kubernetes.io/projected/99b8492d-66ff-496d-8514-5d50a498e04a-kube-api-access-987lm\") pod \"multus-additional-cni-plugins-j9tk4\" (UID: \"99b8492d-66ff-496d-8514-5d50a498e04a\") " pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.975446 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-nrck8" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.976060 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:36Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:36 crc kubenswrapper[4923]: I1009 10:05:36.992007 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:36Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.010490 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.025629 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.043074 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.053852 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-run-netns\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.053894 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-var-lib-cni-multus\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.053912 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-var-lib-kubelet\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.053930 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/1130e920-d4a8-44fe-8bb4-050213d1cff2-rootfs\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.053946 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-var-lib-cni-bin\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.053962 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-run-multus-certs\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.053991 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzq58\" (UniqueName: \"kubernetes.io/projected/1130e920-d4a8-44fe-8bb4-050213d1cff2-kube-api-access-nzq58\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054010 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-system-cni-dir\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054023 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-cnibin\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054036 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khbl2\" (UniqueName: \"kubernetes.io/projected/d5c5f455-4e94-4f9a-b6fa-2535685c6354-kube-api-access-khbl2\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054051 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-cni-dir\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054064 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-os-release\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054078 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d5c5f455-4e94-4f9a-b6fa-2535685c6354-cni-binary-copy\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054096 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1130e920-d4a8-44fe-8bb4-050213d1cff2-mcd-auth-proxy-config\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054121 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-run-k8s-cni-cncf-io\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054134 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-conf-dir\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054153 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1130e920-d4a8-44fe-8bb4-050213d1cff2-proxy-tls\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054172 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-daemon-config\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054185 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-socket-dir-parent\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054199 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-hostroot\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054212 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-etc-kubernetes\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054288 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-etc-kubernetes\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054399 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-run-netns\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054422 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-var-lib-cni-multus\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054442 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-var-lib-kubelet\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054463 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/1130e920-d4a8-44fe-8bb4-050213d1cff2-rootfs\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054483 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-var-lib-cni-bin\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054502 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-run-multus-certs\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054777 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-system-cni-dir\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054817 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-cnibin\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.054974 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-cni-dir\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.055128 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-os-release\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.055886 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d5c5f455-4e94-4f9a-b6fa-2535685c6354-cni-binary-copy\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.056024 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-host-run-k8s-cni-cncf-io\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.056368 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-daemon-config\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.056429 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-socket-dir-parent\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.056463 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-hostroot\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.056503 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d5c5f455-4e94-4f9a-b6fa-2535685c6354-multus-conf-dir\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.056703 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1130e920-d4a8-44fe-8bb4-050213d1cff2-mcd-auth-proxy-config\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.058682 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1130e920-d4a8-44fe-8bb4-050213d1cff2-proxy-tls\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.060548 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.065429 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.069532 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.077495 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khbl2\" (UniqueName: \"kubernetes.io/projected/d5c5f455-4e94-4f9a-b6fa-2535685c6354-kube-api-access-khbl2\") pod \"multus-4cn4f\" (UID: \"d5c5f455-4e94-4f9a-b6fa-2535685c6354\") " pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.077506 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzq58\" (UniqueName: \"kubernetes.io/projected/1130e920-d4a8-44fe-8bb4-050213d1cff2-kube-api-access-nzq58\") pod \"machine-config-daemon-frh4j\" (UID: \"1130e920-d4a8-44fe-8bb4-050213d1cff2\") " pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.086098 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.095185 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-4cn4f" Oct 09 10:05:37 crc kubenswrapper[4923]: W1009 10:05:37.098933 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc2a175c_508c_4f50_8205_013a9482c989.slice/crio-6e6a1e686208e721282523d9ce50dd44ac2b857ab0c9e47061a12222cf97e75f WatchSource:0}: Error finding container 6e6a1e686208e721282523d9ce50dd44ac2b857ab0c9e47061a12222cf97e75f: Status 404 returned error can't find the container with id 6e6a1e686208e721282523d9ce50dd44ac2b857ab0c9e47061a12222cf97e75f Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.099930 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.107425 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.124250 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: W1009 10:05:37.125404 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1130e920_d4a8_44fe_8bb4_050213d1cff2.slice/crio-ef2b62348c027d282a0c5fcbb2e5fbf98120b97ced4baa5cc8d5af2a6051957d WatchSource:0}: Error finding container ef2b62348c027d282a0c5fcbb2e5fbf98120b97ced4baa5cc8d5af2a6051957d: Status 404 returned error can't find the container with id ef2b62348c027d282a0c5fcbb2e5fbf98120b97ced4baa5cc8d5af2a6051957d Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.154658 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.154847 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.154887 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.154912 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.154939 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:05:39.154912349 +0000 UTC m=+25.223094105 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.154949 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.155027 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.155040 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:39.155015752 +0000 UTC m=+25.223197568 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.155040 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.155064 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.155077 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.155067 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:39.155058813 +0000 UTC m=+25.223240569 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.155111 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:39.155100634 +0000 UTC m=+25.223282470 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.256186 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.256418 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.257564 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.257604 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.257670 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:39.257653291 +0000 UTC m=+25.325835047 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.600349 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.600398 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.600478 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.600479 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.600621 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:37 crc kubenswrapper[4923]: E1009 10:05:37.600710 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.745429 4923 generic.go:334] "Generic (PLEG): container finished" podID="99b8492d-66ff-496d-8514-5d50a498e04a" containerID="9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2" exitCode=0 Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.745507 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" event={"ID":"99b8492d-66ff-496d-8514-5d50a498e04a","Type":"ContainerDied","Data":"9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.745556 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" event={"ID":"99b8492d-66ff-496d-8514-5d50a498e04a","Type":"ContainerStarted","Data":"29107e51bb57bfd2a977a9e5c4d08af8e70b2c1fc014e2f2e374bc9eba937994"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.748948 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b" exitCode=0 Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.749011 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.749039 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"6e6a1e686208e721282523d9ce50dd44ac2b857ab0c9e47061a12222cf97e75f"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.751111 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.751167 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.751182 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"ef2b62348c027d282a0c5fcbb2e5fbf98120b97ced4baa5cc8d5af2a6051957d"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.753018 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nrck8" event={"ID":"7fd8165c-555f-4ded-8081-cb273d195245","Type":"ContainerStarted","Data":"11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.753045 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-nrck8" event={"ID":"7fd8165c-555f-4ded-8081-cb273d195245","Type":"ContainerStarted","Data":"ab8da493b0152514987c299d6d5107dc9ae13eb1be69a397794c4e19fd728a0a"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.755300 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4cn4f" event={"ID":"d5c5f455-4e94-4f9a-b6fa-2535685c6354","Type":"ContainerStarted","Data":"c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.755331 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4cn4f" event={"ID":"d5c5f455-4e94-4f9a-b6fa-2535685c6354","Type":"ContainerStarted","Data":"fbb2bdb5fe8847cf2b72f190d2569d8438c12b061d45d5ed56232a67cd101ebe"} Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.770116 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.794429 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.816668 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.832172 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.846095 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.872948 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.914305 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.939326 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:37 crc kubenswrapper[4923]: I1009 10:05:37.964948 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.001806 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:37Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.017719 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.032240 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.043695 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.056369 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.077971 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.102069 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.115000 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.128192 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.142884 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.157942 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.176579 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.189529 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.205801 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.221576 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.236002 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.246411 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.760308 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.760653 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.760671 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.760683 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.760693 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.761501 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f"} Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.764004 4923 generic.go:334] "Generic (PLEG): container finished" podID="99b8492d-66ff-496d-8514-5d50a498e04a" containerID="ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0" exitCode=0 Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.764043 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" event={"ID":"99b8492d-66ff-496d-8514-5d50a498e04a","Type":"ContainerDied","Data":"ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0"} Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.776651 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.792689 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.806504 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.822741 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.844156 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.859112 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.874094 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.887426 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.900333 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.915409 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.927958 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.938469 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.950482 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.962827 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.979103 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:38 crc kubenswrapper[4923]: I1009 10:05:38.999101 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.017849 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.031495 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.050121 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.063592 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.077343 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.090808 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.105009 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.118226 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.130711 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.140180 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.174454 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.174551 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.174625 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:05:43.17459727 +0000 UTC m=+29.242779026 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.174639 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.174671 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.174694 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.174729 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:43.174693293 +0000 UTC m=+29.242875139 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.174804 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.174815 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.174824 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.174851 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:43.174843097 +0000 UTC m=+29.243024853 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.174892 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.174912 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:43.174906239 +0000 UTC m=+29.243087995 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.275173 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.275375 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.275407 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.275417 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.275476 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:43.275457499 +0000 UTC m=+29.343639245 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.601052 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.601104 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.601186 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.601257 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.601486 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:39 crc kubenswrapper[4923]: E1009 10:05:39.601670 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.718630 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-k72w6"] Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.719006 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.721113 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.722307 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.722821 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.724153 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.737216 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.751140 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.768247 4923 generic.go:334] "Generic (PLEG): container finished" podID="99b8492d-66ff-496d-8514-5d50a498e04a" containerID="069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f" exitCode=0 Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.768326 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" event={"ID":"99b8492d-66ff-496d-8514-5d50a498e04a","Type":"ContainerDied","Data":"069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f"} Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.772671 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.779846 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szxlf\" (UniqueName: \"kubernetes.io/projected/58586169-cd8e-4003-a38e-79c600a6a845-kube-api-access-szxlf\") pod \"node-ca-k72w6\" (UID: \"58586169-cd8e-4003-a38e-79c600a6a845\") " pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.779898 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/58586169-cd8e-4003-a38e-79c600a6a845-serviceca\") pod \"node-ca-k72w6\" (UID: \"58586169-cd8e-4003-a38e-79c600a6a845\") " pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.779957 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58586169-cd8e-4003-a38e-79c600a6a845-host\") pod \"node-ca-k72w6\" (UID: \"58586169-cd8e-4003-a38e-79c600a6a845\") " pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.796613 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.819773 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.838180 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.857152 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.871274 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.881081 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szxlf\" (UniqueName: \"kubernetes.io/projected/58586169-cd8e-4003-a38e-79c600a6a845-kube-api-access-szxlf\") pod \"node-ca-k72w6\" (UID: \"58586169-cd8e-4003-a38e-79c600a6a845\") " pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.881134 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/58586169-cd8e-4003-a38e-79c600a6a845-serviceca\") pod \"node-ca-k72w6\" (UID: \"58586169-cd8e-4003-a38e-79c600a6a845\") " pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.881169 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58586169-cd8e-4003-a38e-79c600a6a845-host\") pod \"node-ca-k72w6\" (UID: \"58586169-cd8e-4003-a38e-79c600a6a845\") " pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.881244 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/58586169-cd8e-4003-a38e-79c600a6a845-host\") pod \"node-ca-k72w6\" (UID: \"58586169-cd8e-4003-a38e-79c600a6a845\") " pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.882207 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/58586169-cd8e-4003-a38e-79c600a6a845-serviceca\") pod \"node-ca-k72w6\" (UID: \"58586169-cd8e-4003-a38e-79c600a6a845\") " pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.884594 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.896179 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.900822 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szxlf\" (UniqueName: \"kubernetes.io/projected/58586169-cd8e-4003-a38e-79c600a6a845-kube-api-access-szxlf\") pod \"node-ca-k72w6\" (UID: \"58586169-cd8e-4003-a38e-79c600a6a845\") " pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.916109 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.933610 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.948261 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.962925 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.978585 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:39 crc kubenswrapper[4923]: I1009 10:05:39.991604 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:39Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.006104 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.019504 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.031919 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.034001 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-k72w6" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.045670 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.056892 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.072103 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.085556 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.095237 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: W1009 10:05:40.098215 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58586169_cd8e_4003_a38e_79c600a6a845.slice/crio-ee866db8d9294ece7504a4077c068e40c97276ecab6847f6aba3cc87d4b4bb64 WatchSource:0}: Error finding container ee866db8d9294ece7504a4077c068e40c97276ecab6847f6aba3cc87d4b4bb64: Status 404 returned error can't find the container with id ee866db8d9294ece7504a4077c068e40c97276ecab6847f6aba3cc87d4b4bb64 Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.109915 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.135054 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.153856 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.168910 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.182219 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.777336 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-k72w6" event={"ID":"58586169-cd8e-4003-a38e-79c600a6a845","Type":"ContainerStarted","Data":"8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc"} Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.777600 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-k72w6" event={"ID":"58586169-cd8e-4003-a38e-79c600a6a845","Type":"ContainerStarted","Data":"ee866db8d9294ece7504a4077c068e40c97276ecab6847f6aba3cc87d4b4bb64"} Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.780127 4923 generic.go:334] "Generic (PLEG): container finished" podID="99b8492d-66ff-496d-8514-5d50a498e04a" containerID="d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990" exitCode=0 Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.780158 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" event={"ID":"99b8492d-66ff-496d-8514-5d50a498e04a","Type":"ContainerDied","Data":"d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990"} Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.794274 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.808850 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.822234 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.837212 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.846373 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.861080 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.873226 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.889086 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.902234 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.911838 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.923929 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.936984 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.948017 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.961376 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.976716 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:40 crc kubenswrapper[4923]: I1009 10:05:40.989018 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.000110 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:40Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.011991 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.021196 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.031885 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.043801 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.056813 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.073957 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.089958 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.102771 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.116981 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.131323 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.145579 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.294402 4923 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.296106 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.296146 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.296162 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.296254 4923 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.304178 4923 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.304415 4923 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.305453 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.305506 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.305520 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.305538 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.305551 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: E1009 10:05:41.317501 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.321210 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.321249 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.321260 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.321277 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.321289 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: E1009 10:05:41.332912 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.336630 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.336670 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.336681 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.336699 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.336709 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: E1009 10:05:41.347172 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.350322 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.350358 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.350366 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.350383 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.350397 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: E1009 10:05:41.361611 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.365217 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.365258 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.365269 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.365284 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.365294 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: E1009 10:05:41.376455 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: E1009 10:05:41.376577 4923 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.378200 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.378236 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.378247 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.378264 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.378274 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.480650 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.480684 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.480701 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.480718 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.480729 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.582963 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.583001 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.583011 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.583029 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.583039 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.600546 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.600564 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:41 crc kubenswrapper[4923]: E1009 10:05:41.600692 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:41 crc kubenswrapper[4923]: E1009 10:05:41.600820 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.600581 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:41 crc kubenswrapper[4923]: E1009 10:05:41.600925 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.684893 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.684934 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.684942 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.684958 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.684968 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.785360 4923 generic.go:334] "Generic (PLEG): container finished" podID="99b8492d-66ff-496d-8514-5d50a498e04a" containerID="a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5" exitCode=0 Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.785413 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" event={"ID":"99b8492d-66ff-496d-8514-5d50a498e04a","Type":"ContainerDied","Data":"a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5"} Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.786297 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.786337 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.786348 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.786367 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.786379 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.798282 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.806184 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.819735 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.837293 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.851828 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.863626 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.875228 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.887642 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.889207 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.889267 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.889281 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.889304 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.889322 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.897741 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.910459 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.925863 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.938601 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.948949 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.962860 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.975201 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:41Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.992200 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.992245 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.992255 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.992271 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:41 crc kubenswrapper[4923]: I1009 10:05:41.992281 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:41Z","lastTransitionTime":"2025-10-09T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.095817 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.095863 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.095874 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.095890 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.095901 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:42Z","lastTransitionTime":"2025-10-09T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.198093 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.198135 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.198146 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.198162 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.198174 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:42Z","lastTransitionTime":"2025-10-09T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.301137 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.301183 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.301195 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.301213 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.301227 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:42Z","lastTransitionTime":"2025-10-09T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.403186 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.403235 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.403246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.403262 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.403272 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:42Z","lastTransitionTime":"2025-10-09T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.505462 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.505512 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.505523 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.505543 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.505557 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:42Z","lastTransitionTime":"2025-10-09T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.607345 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.607390 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.607402 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.607416 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.607430 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:42Z","lastTransitionTime":"2025-10-09T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.709705 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.709747 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.709776 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.709795 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.709806 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:42Z","lastTransitionTime":"2025-10-09T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.803804 4923 generic.go:334] "Generic (PLEG): container finished" podID="99b8492d-66ff-496d-8514-5d50a498e04a" containerID="6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73" exitCode=0 Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.803842 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" event={"ID":"99b8492d-66ff-496d-8514-5d50a498e04a","Type":"ContainerDied","Data":"6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.812188 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.812225 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.812239 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.812258 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.812269 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:42Z","lastTransitionTime":"2025-10-09T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.819733 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.831627 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.850455 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.863240 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.874219 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.890320 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.909311 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.914395 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.914433 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.914449 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.914469 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.914480 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:42Z","lastTransitionTime":"2025-10-09T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.924201 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.935774 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.947673 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.960952 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.973294 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.984595 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:42 crc kubenswrapper[4923]: I1009 10:05:42.997237 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:42Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.017166 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.017215 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.017225 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.017240 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.017250 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.119560 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.119590 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.119600 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.119615 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.119625 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.206611 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.206739 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.206805 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.206829 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.206945 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.206965 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.206985 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.206990 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:05:51.206967149 +0000 UTC m=+37.275148905 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.206996 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.207009 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:51.20700227 +0000 UTC m=+37.275184026 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.207033 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.207055 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:51.207032591 +0000 UTC m=+37.275214367 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.207091 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:51.207079272 +0000 UTC m=+37.275261028 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.221720 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.221805 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.221830 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.221848 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.221860 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.307176 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.307366 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.307530 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.307550 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.307616 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 10:05:51.307595992 +0000 UTC m=+37.375777748 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.324169 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.324205 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.324214 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.324230 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.324239 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.426951 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.426991 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.427000 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.427018 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.427028 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.528964 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.529003 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.529013 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.529027 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.529038 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.600554 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.600598 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.600664 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.600691 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.600796 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:43 crc kubenswrapper[4923]: E1009 10:05:43.600865 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.631497 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.631533 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.631541 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.631559 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.631569 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.733638 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.733669 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.733677 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.733690 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.733699 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.809500 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" event={"ID":"99b8492d-66ff-496d-8514-5d50a498e04a","Type":"ContainerStarted","Data":"e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.813824 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.814211 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.814234 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.814246 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.823956 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.836331 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.836374 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.836387 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.836403 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.836413 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.836865 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.847848 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.857572 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.860051 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.860880 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.874896 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.885778 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.894709 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.930207 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.939208 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.939245 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.939254 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.939268 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.939278 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:43Z","lastTransitionTime":"2025-10-09T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.941934 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.952662 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.963964 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.978958 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:43 crc kubenswrapper[4923]: I1009 10:05:43.992550 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:43Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.002735 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.012387 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.022006 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.033333 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.041972 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.042011 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.042024 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.042043 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.042054 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.050536 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.062922 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.074814 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.084955 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.098182 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.110507 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.124369 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.138105 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.144433 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.144476 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.144485 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.144498 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.144508 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.149070 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.161426 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.172660 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.246926 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.246960 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.246968 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.246981 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.246991 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.349140 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.349222 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.349242 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.349259 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.349272 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.451925 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.451966 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.451990 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.452011 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.452025 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.554437 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.554683 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.554695 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.554712 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.554731 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.614201 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.628022 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.641989 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.656581 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.656628 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.656637 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.656650 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.656659 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.657103 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.671743 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.691863 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.704466 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.716941 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.727998 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.741107 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.753596 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.758223 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.758251 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.758260 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.758274 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.758282 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.769976 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.787911 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.799705 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:44Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.860147 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.860184 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.860194 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.860209 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.860221 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.962511 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.962562 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.962574 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.962589 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:44 crc kubenswrapper[4923]: I1009 10:05:44.962600 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:44Z","lastTransitionTime":"2025-10-09T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.065208 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.065519 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.065529 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.065544 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.065555 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.167278 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.167308 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.167317 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.167330 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.167339 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.273119 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.273200 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.273211 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.273246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.273259 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.375951 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.375987 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.375998 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.376013 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.376024 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.478910 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.478953 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.478964 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.478979 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.478990 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.581827 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.581878 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.581886 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.581899 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.581927 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.601098 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.601155 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:45 crc kubenswrapper[4923]: E1009 10:05:45.601224 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:45 crc kubenswrapper[4923]: E1009 10:05:45.601285 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.601352 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:45 crc kubenswrapper[4923]: E1009 10:05:45.601399 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.684359 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.684405 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.684415 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.684431 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.684439 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.786820 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.786866 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.786876 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.786893 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.786904 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.889585 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.889652 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.889902 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.889929 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.889938 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.992798 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.992830 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.992840 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.992856 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:45 crc kubenswrapper[4923]: I1009 10:05:45.992867 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:45Z","lastTransitionTime":"2025-10-09T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.095246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.095274 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.095282 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.095295 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.095303 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:46Z","lastTransitionTime":"2025-10-09T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.197505 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.197537 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.197548 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.197561 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.197569 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:46Z","lastTransitionTime":"2025-10-09T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.300608 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.300647 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.300658 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.300675 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.300686 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:46Z","lastTransitionTime":"2025-10-09T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.403719 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.403774 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.403788 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.403804 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.403816 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:46Z","lastTransitionTime":"2025-10-09T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.506007 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.506049 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.506063 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.506083 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.506096 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:46Z","lastTransitionTime":"2025-10-09T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.608045 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.608096 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.608113 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.608136 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.608152 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:46Z","lastTransitionTime":"2025-10-09T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.710863 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.710920 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.710937 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.710979 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.710997 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:46Z","lastTransitionTime":"2025-10-09T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.813102 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.813164 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.813174 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.813189 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.813199 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:46Z","lastTransitionTime":"2025-10-09T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.824016 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/0.log" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.826461 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2" exitCode=1 Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.826516 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.827323 4923 scope.go:117] "RemoveContainer" containerID="74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.842855 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:46Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.856405 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:46Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.874624 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:46Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.898656 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:46Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.919036 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.919076 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.919085 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.919100 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.919109 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:46Z","lastTransitionTime":"2025-10-09T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.922468 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:46Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.939988 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:46Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.956914 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:46Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.978925 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:46Z\\\",\\\"message\\\":\\\" (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395161 6198 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395383 6198 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 10:05:46.395588 6198 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395812 6198 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 10:05:46.396079 6198 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1009 10:05:46.396090 6198 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1009 10:05:46.396134 6198 factory.go:656] Stopping watch factory\\\\nI1009 10:05:46.396149 6198 ovnkube.go:599] Stopped ovnkube\\\\nI1009 10:05:46.396146 6198 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1009 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:46Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:46 crc kubenswrapper[4923]: I1009 10:05:46.993850 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:46Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.005784 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.017013 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.021327 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.021368 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.021378 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.021394 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.021405 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.031127 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.042686 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.056440 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.127185 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.127687 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.127700 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.127720 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.127733 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.230413 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.230448 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.230457 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.230471 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.230479 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.333423 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.333458 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.333468 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.333483 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.333493 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.435106 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.435143 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.435151 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.435165 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.435173 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.537779 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.537826 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.537840 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.537858 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.537887 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.601246 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:47 crc kubenswrapper[4923]: E1009 10:05:47.601368 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.601256 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:47 crc kubenswrapper[4923]: E1009 10:05:47.601513 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.601697 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:47 crc kubenswrapper[4923]: E1009 10:05:47.601781 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.639653 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.639709 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.639726 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.639743 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.639771 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.741681 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.741721 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.741731 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.741764 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.741777 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.831162 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/1.log" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.831729 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/0.log" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.834345 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8" exitCode=1 Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.834383 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.834439 4923 scope.go:117] "RemoveContainer" containerID="74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.835019 4923 scope.go:117] "RemoveContainer" containerID="c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8" Oct 09 10:05:47 crc kubenswrapper[4923]: E1009 10:05:47.835244 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\"" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.844059 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.844088 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.844096 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.844109 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.844118 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.850602 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.867967 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.880146 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.893268 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.906993 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.922057 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.936561 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.946435 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.946470 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.946481 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.946500 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.946518 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:47Z","lastTransitionTime":"2025-10-09T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.956448 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:46Z\\\",\\\"message\\\":\\\" (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395161 6198 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395383 6198 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 10:05:46.395588 6198 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395812 6198 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 10:05:46.396079 6198 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1009 10:05:46.396090 6198 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1009 10:05:46.396134 6198 factory.go:656] Stopping watch factory\\\\nI1009 10:05:46.396149 6198 ovnkube.go:599] Stopped ovnkube\\\\nI1009 10:05:46.396146 6198 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1009 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:47Z\\\",\\\"message\\\":\\\"16 6322 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1009 10:05:47.775227 6322 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-frh4j in node crc\\\\nF1009 10:05:47.775227 6322 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 10:05:47.775228 6322 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1009 10:05:47.775241 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.971393 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:47 crc kubenswrapper[4923]: I1009 10:05:47.987186 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.000184 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.013270 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.025275 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.040023 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.048411 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.048448 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.048457 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.048470 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.048483 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.151217 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.151267 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.151280 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.151300 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.151311 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.253675 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.253704 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.253714 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.253726 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.253736 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.356027 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.356088 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.356100 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.356117 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.356130 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.460113 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.460161 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.460173 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.460196 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.460214 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.562980 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.563034 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.563044 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.563056 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.563064 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.666233 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.666292 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.666306 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.666328 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.666341 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.769073 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.769125 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.769135 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.769156 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.769169 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.837824 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56"] Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.838398 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.840909 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/1.log" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.842496 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.842809 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.857125 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.868978 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea1bca1f-8564-4e84-b48e-b105ed075b44-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.869223 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea1bca1f-8564-4e84-b48e-b105ed075b44-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.869332 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea1bca1f-8564-4e84-b48e-b105ed075b44-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.869450 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9r9m4\" (UniqueName: \"kubernetes.io/projected/ea1bca1f-8564-4e84-b48e-b105ed075b44-kube-api-access-9r9m4\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.870167 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.871997 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.872060 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.872075 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.872100 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.872115 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.886860 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.899742 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.927742 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:46Z\\\",\\\"message\\\":\\\" (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395161 6198 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395383 6198 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 10:05:46.395588 6198 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395812 6198 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 10:05:46.396079 6198 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1009 10:05:46.396090 6198 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1009 10:05:46.396134 6198 factory.go:656] Stopping watch factory\\\\nI1009 10:05:46.396149 6198 ovnkube.go:599] Stopped ovnkube\\\\nI1009 10:05:46.396146 6198 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1009 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:47Z\\\",\\\"message\\\":\\\"16 6322 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1009 10:05:47.775227 6322 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-frh4j in node crc\\\\nF1009 10:05:47.775227 6322 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 10:05:47.775228 6322 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1009 10:05:47.775241 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.944997 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.963374 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.970649 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea1bca1f-8564-4e84-b48e-b105ed075b44-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.970727 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea1bca1f-8564-4e84-b48e-b105ed075b44-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.970777 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea1bca1f-8564-4e84-b48e-b105ed075b44-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.970826 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9r9m4\" (UniqueName: \"kubernetes.io/projected/ea1bca1f-8564-4e84-b48e-b105ed075b44-kube-api-access-9r9m4\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.971530 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ea1bca1f-8564-4e84-b48e-b105ed075b44-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.972275 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ea1bca1f-8564-4e84-b48e-b105ed075b44-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.976046 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.976082 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.976096 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.976113 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.976125 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:48Z","lastTransitionTime":"2025-10-09T10:05:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.979396 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.982026 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ea1bca1f-8564-4e84-b48e-b105ed075b44-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.994988 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:48Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:48 crc kubenswrapper[4923]: I1009 10:05:48.995244 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9r9m4\" (UniqueName: \"kubernetes.io/projected/ea1bca1f-8564-4e84-b48e-b105ed075b44-kube-api-access-9r9m4\") pod \"ovnkube-control-plane-749d76644c-bxn56\" (UID: \"ea1bca1f-8564-4e84-b48e-b105ed075b44\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.015206 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.029871 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.044692 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.059589 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.074016 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.079115 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.079167 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.079184 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.079206 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.079217 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:49Z","lastTransitionTime":"2025-10-09T10:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.088409 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.153359 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" Oct 09 10:05:49 crc kubenswrapper[4923]: W1009 10:05:49.165405 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podea1bca1f_8564_4e84_b48e_b105ed075b44.slice/crio-4e96f1c9d414ad02c9a62f53aec2ff3f900b9c2dd89c0b80d858d56fc31b9b07 WatchSource:0}: Error finding container 4e96f1c9d414ad02c9a62f53aec2ff3f900b9c2dd89c0b80d858d56fc31b9b07: Status 404 returned error can't find the container with id 4e96f1c9d414ad02c9a62f53aec2ff3f900b9c2dd89c0b80d858d56fc31b9b07 Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.181456 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.181485 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.181497 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.181511 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.181521 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:49Z","lastTransitionTime":"2025-10-09T10:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.284308 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.284356 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.284368 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.284385 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.284396 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:49Z","lastTransitionTime":"2025-10-09T10:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.390171 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.390359 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.390421 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.390631 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.390644 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:49Z","lastTransitionTime":"2025-10-09T10:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.494123 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.494174 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.494187 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.494207 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.494220 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:49Z","lastTransitionTime":"2025-10-09T10:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.596326 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.596370 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.596380 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.596395 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.596404 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:49Z","lastTransitionTime":"2025-10-09T10:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.600685 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.600834 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:49 crc kubenswrapper[4923]: E1009 10:05:49.600995 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:49 crc kubenswrapper[4923]: E1009 10:05:49.601148 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.601390 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:49 crc kubenswrapper[4923]: E1009 10:05:49.601585 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.698639 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.698696 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.698709 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.698731 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.698745 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:49Z","lastTransitionTime":"2025-10-09T10:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.801973 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.802016 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.802028 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.802046 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.802059 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:49Z","lastTransitionTime":"2025-10-09T10:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.852146 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" event={"ID":"ea1bca1f-8564-4e84-b48e-b105ed075b44","Type":"ContainerStarted","Data":"60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.852215 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" event={"ID":"ea1bca1f-8564-4e84-b48e-b105ed075b44","Type":"ContainerStarted","Data":"e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.852226 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" event={"ID":"ea1bca1f-8564-4e84-b48e-b105ed075b44","Type":"ContainerStarted","Data":"4e96f1c9d414ad02c9a62f53aec2ff3f900b9c2dd89c0b80d858d56fc31b9b07"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.867869 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.880919 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.900714 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.904377 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.904441 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.904456 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.904473 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.904511 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:49Z","lastTransitionTime":"2025-10-09T10:05:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.914083 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.926030 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.937984 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.949018 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.963506 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.974647 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.984067 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:49 crc kubenswrapper[4923]: I1009 10:05:49.994566 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:49Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.006970 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.007019 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.007027 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.007040 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.007050 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.013616 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:46Z\\\",\\\"message\\\":\\\" (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395161 6198 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395383 6198 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 10:05:46.395588 6198 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395812 6198 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 10:05:46.396079 6198 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1009 10:05:46.396090 6198 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1009 10:05:46.396134 6198 factory.go:656] Stopping watch factory\\\\nI1009 10:05:46.396149 6198 ovnkube.go:599] Stopped ovnkube\\\\nI1009 10:05:46.396146 6198 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1009 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:47Z\\\",\\\"message\\\":\\\"16 6322 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1009 10:05:47.775227 6322 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-frh4j in node crc\\\\nF1009 10:05:47.775227 6322 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 10:05:47.775228 6322 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1009 10:05:47.775241 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.027380 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.039093 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.050337 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.109661 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.109710 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.109721 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.109739 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.109773 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.212548 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.212591 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.212602 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.212616 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.212625 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.315546 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.315586 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.315595 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.315610 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.315620 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.417567 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.417625 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.417635 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.417651 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.417663 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.520417 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.520472 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.520482 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.520497 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.520506 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.622425 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.622464 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.622479 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.622494 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.622503 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.673670 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-dxnsb"] Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.674138 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:50 crc kubenswrapper[4923]: E1009 10:05:50.674192 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.686649 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.698615 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.709245 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.723347 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.727267 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.727321 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.727333 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.727349 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.727363 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.738201 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.750456 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.761384 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.771268 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.783498 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.791237 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.791281 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgphv\" (UniqueName: \"kubernetes.io/projected/0ffdd04f-94e8-4fca-9676-ef87952debcc-kube-api-access-pgphv\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.793340 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.804908 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.817665 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.829958 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.830001 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.830015 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.830030 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.830042 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.836927 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:46Z\\\",\\\"message\\\":\\\" (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395161 6198 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395383 6198 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 10:05:46.395588 6198 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395812 6198 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 10:05:46.396079 6198 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1009 10:05:46.396090 6198 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1009 10:05:46.396134 6198 factory.go:656] Stopping watch factory\\\\nI1009 10:05:46.396149 6198 ovnkube.go:599] Stopped ovnkube\\\\nI1009 10:05:46.396146 6198 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1009 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:47Z\\\",\\\"message\\\":\\\"16 6322 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1009 10:05:47.775227 6322 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-frh4j in node crc\\\\nF1009 10:05:47.775227 6322 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 10:05:47.775228 6322 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1009 10:05:47.775241 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.849411 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.860632 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.871345 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:50Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.891928 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.891981 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgphv\" (UniqueName: \"kubernetes.io/projected/0ffdd04f-94e8-4fca-9676-ef87952debcc-kube-api-access-pgphv\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:50 crc kubenswrapper[4923]: E1009 10:05:50.892132 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:50 crc kubenswrapper[4923]: E1009 10:05:50.892228 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs podName:0ffdd04f-94e8-4fca-9676-ef87952debcc nodeName:}" failed. No retries permitted until 2025-10-09 10:05:51.392205032 +0000 UTC m=+37.460386858 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs") pod "network-metrics-daemon-dxnsb" (UID: "0ffdd04f-94e8-4fca-9676-ef87952debcc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.907451 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgphv\" (UniqueName: \"kubernetes.io/projected/0ffdd04f-94e8-4fca-9676-ef87952debcc-kube-api-access-pgphv\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.932528 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.932566 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.932574 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.932590 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:50 crc kubenswrapper[4923]: I1009 10:05:50.932601 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:50Z","lastTransitionTime":"2025-10-09T10:05:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.034962 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.035002 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.035016 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.035031 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.035040 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.137133 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.137162 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.137170 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.137182 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.137191 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.239470 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.239502 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.239510 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.239523 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.239531 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.294218 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.294322 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.294347 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.294365 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.294446 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:06:07.294417304 +0000 UTC m=+53.362599060 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.294478 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.294492 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.294501 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.294519 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.294573 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.294541 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 10:06:07.294528447 +0000 UTC m=+53.362710203 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.294652 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:06:07.29462589 +0000 UTC m=+53.362807706 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.294672 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:06:07.294661961 +0000 UTC m=+53.362843817 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.341524 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.341563 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.341573 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.341587 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.341595 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.395366 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.395632 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.395563 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.395837 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.395856 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.395869 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.395917 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 10:06:07.39590049 +0000 UTC m=+53.464082246 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.396372 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs podName:0ffdd04f-94e8-4fca-9676-ef87952debcc nodeName:}" failed. No retries permitted until 2025-10-09 10:05:52.396329973 +0000 UTC m=+38.464511739 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs") pod "network-metrics-daemon-dxnsb" (UID: "0ffdd04f-94e8-4fca-9676-ef87952debcc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.443701 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.444055 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.444066 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.444085 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.444098 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.546526 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.546574 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.546585 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.546610 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.546619 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.601297 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.601367 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.601427 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.601443 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.601490 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.601584 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.649070 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.649103 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.649116 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.649130 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.649139 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.650019 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.650050 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.650061 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.650074 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.650084 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.662220 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:51Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.665618 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.665679 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.665687 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.665700 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.665709 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.676458 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:51Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.679868 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.679906 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.679918 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.679936 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.679950 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.690988 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:51Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.694064 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.694100 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.694112 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.694128 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.694139 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.705647 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:51Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.708962 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.709014 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.709027 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.709045 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.709059 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.721374 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:51Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:51 crc kubenswrapper[4923]: E1009 10:05:51.721543 4923 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.751658 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.751696 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.751707 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.751723 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.751735 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.853888 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.853938 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.853950 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.853966 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.853978 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.955994 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.956057 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.956068 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.956082 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:51 crc kubenswrapper[4923]: I1009 10:05:51.956091 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:51Z","lastTransitionTime":"2025-10-09T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.059501 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.059557 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.059567 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.059583 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.059593 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.162563 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.162621 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.162632 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.162651 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.162665 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.265943 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.266022 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.266043 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.266066 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.266082 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.368836 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.368876 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.368886 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.368902 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.368912 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.406001 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:52 crc kubenswrapper[4923]: E1009 10:05:52.406161 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:52 crc kubenswrapper[4923]: E1009 10:05:52.406230 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs podName:0ffdd04f-94e8-4fca-9676-ef87952debcc nodeName:}" failed. No retries permitted until 2025-10-09 10:05:54.406211989 +0000 UTC m=+40.474393765 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs") pod "network-metrics-daemon-dxnsb" (UID: "0ffdd04f-94e8-4fca-9676-ef87952debcc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.470492 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.470811 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.470884 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.470956 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.471015 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.573469 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.573523 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.573537 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.573556 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.573571 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.601055 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:52 crc kubenswrapper[4923]: E1009 10:05:52.601202 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.675783 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.675821 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.675832 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.675847 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.675857 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.778789 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.778827 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.778837 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.778854 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.778862 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.880681 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.880722 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.880730 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.880744 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.880773 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.983113 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.983154 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.983166 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.983187 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:52 crc kubenswrapper[4923]: I1009 10:05:52.983202 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:52Z","lastTransitionTime":"2025-10-09T10:05:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.085794 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.085839 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.085857 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.085873 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.085885 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:53Z","lastTransitionTime":"2025-10-09T10:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.188060 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.188098 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.188110 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.188124 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.188136 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:53Z","lastTransitionTime":"2025-10-09T10:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.290649 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.290788 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.290817 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.290842 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.291370 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:53Z","lastTransitionTime":"2025-10-09T10:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.394463 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.394532 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.394545 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.394562 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.394599 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:53Z","lastTransitionTime":"2025-10-09T10:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.497168 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.497208 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.497218 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.497233 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.497244 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:53Z","lastTransitionTime":"2025-10-09T10:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.600139 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.600174 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.600220 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.600223 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:53 crc kubenswrapper[4923]: E1009 10:05:53.600361 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.600188 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:53 crc kubenswrapper[4923]: E1009 10:05:53.600433 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.600480 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:53 crc kubenswrapper[4923]: E1009 10:05:53.600532 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.600235 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.600627 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:53Z","lastTransitionTime":"2025-10-09T10:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.704378 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.704436 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.704450 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.704469 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.704481 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:53Z","lastTransitionTime":"2025-10-09T10:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.807125 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.807173 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.807184 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.807203 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.807215 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:53Z","lastTransitionTime":"2025-10-09T10:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.910092 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.910127 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.910136 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.910150 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:53 crc kubenswrapper[4923]: I1009 10:05:53.910161 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:53Z","lastTransitionTime":"2025-10-09T10:05:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.013789 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.013841 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.013853 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.013877 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.013893 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.116673 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.116716 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.116725 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.116738 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.116769 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.220031 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.220079 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.220094 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.220117 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.220131 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.323216 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.323264 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.323276 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.323295 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.323311 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.426034 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.426095 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.426112 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.426134 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.426148 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.429702 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:54 crc kubenswrapper[4923]: E1009 10:05:54.429947 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:54 crc kubenswrapper[4923]: E1009 10:05:54.430057 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs podName:0ffdd04f-94e8-4fca-9676-ef87952debcc nodeName:}" failed. No retries permitted until 2025-10-09 10:05:58.430025924 +0000 UTC m=+44.498207851 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs") pod "network-metrics-daemon-dxnsb" (UID: "0ffdd04f-94e8-4fca-9676-ef87952debcc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.529638 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.529707 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.529721 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.529743 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.529777 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.601087 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:54 crc kubenswrapper[4923]: E1009 10:05:54.601286 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.619799 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.631960 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.632034 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.632050 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.632066 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.632078 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.632325 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.650988 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.675367 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:46Z\\\",\\\"message\\\":\\\" (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395161 6198 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395383 6198 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 10:05:46.395588 6198 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395812 6198 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 10:05:46.396079 6198 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1009 10:05:46.396090 6198 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1009 10:05:46.396134 6198 factory.go:656] Stopping watch factory\\\\nI1009 10:05:46.396149 6198 ovnkube.go:599] Stopped ovnkube\\\\nI1009 10:05:46.396146 6198 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1009 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:47Z\\\",\\\"message\\\":\\\"16 6322 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1009 10:05:47.775227 6322 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-frh4j in node crc\\\\nF1009 10:05:47.775227 6322 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 10:05:47.775228 6322 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1009 10:05:47.775241 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.698902 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.713180 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.724261 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.734141 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.734169 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.734179 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.734194 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.734206 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.738065 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.751585 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.763309 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.781717 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.795389 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.807010 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.820372 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.831656 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.836455 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.836495 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.836505 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.836520 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.836529 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.848234 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:54Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.938814 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.938860 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.938943 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.938964 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:54 crc kubenswrapper[4923]: I1009 10:05:54.938977 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:54Z","lastTransitionTime":"2025-10-09T10:05:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.041337 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.041383 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.041391 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.041405 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.041417 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.144182 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.144224 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.144232 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.144246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.144255 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.242920 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.246511 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.246553 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.246591 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.246607 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.246620 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.256995 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.272425 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.293015 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://74723b7f7b33e041a10d6a7dea04dbfb5a64f6fcabc89c1537ea71ef68e555b2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:46Z\\\",\\\"message\\\":\\\" (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395161 6198 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395383 6198 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 10:05:46.395588 6198 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 10:05:46.395812 6198 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 10:05:46.396079 6198 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1009 10:05:46.396090 6198 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1009 10:05:46.396134 6198 factory.go:656] Stopping watch factory\\\\nI1009 10:05:46.396149 6198 ovnkube.go:599] Stopped ovnkube\\\\nI1009 10:05:46.396146 6198 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1009 10\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:47Z\\\",\\\"message\\\":\\\"16 6322 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1009 10:05:47.775227 6322 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-frh4j in node crc\\\\nF1009 10:05:47.775227 6322 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 10:05:47.775228 6322 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1009 10:05:47.775241 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.309158 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.322792 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.335659 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.349148 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.349186 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.349195 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.349212 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.349221 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.352085 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.372237 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.386488 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.398927 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.409646 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.422553 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.435668 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.446913 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.451726 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.451798 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.451814 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.451835 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.451846 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.460118 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.471713 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:55Z is after 2025-08-24T17:21:41Z" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.553914 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.554213 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.554290 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.554378 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.554495 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.600811 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.601232 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:55 crc kubenswrapper[4923]: E1009 10:05:55.601392 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.601581 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:55 crc kubenswrapper[4923]: E1009 10:05:55.601849 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:55 crc kubenswrapper[4923]: E1009 10:05:55.601924 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.656992 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.657051 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.657066 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.657082 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.657093 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.759095 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.759184 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.759199 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.759215 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.759227 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.862044 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.862075 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.862084 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.862098 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.862112 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.964329 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.964377 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.964386 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.964401 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:55 crc kubenswrapper[4923]: I1009 10:05:55.964414 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:55Z","lastTransitionTime":"2025-10-09T10:05:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.069052 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.069107 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.069118 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.069142 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.069156 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.171331 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.171378 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.171387 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.171402 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.171412 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.274619 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.274662 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.274671 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.274689 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.274699 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.377906 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.378194 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.378265 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.378347 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.378413 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.482430 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.482484 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.482496 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.482515 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.482528 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.584909 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.584944 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.584952 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.584965 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.584974 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.601591 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:56 crc kubenswrapper[4923]: E1009 10:05:56.601747 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.686938 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.686981 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.686990 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.687007 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.687017 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.790410 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.790468 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.790479 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.790499 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.790510 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.893693 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.893785 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.893797 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.893812 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.893822 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.996731 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.996809 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.996823 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.996841 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:56 crc kubenswrapper[4923]: I1009 10:05:56.996855 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:56Z","lastTransitionTime":"2025-10-09T10:05:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.099221 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.099253 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.099261 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.099274 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.099284 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:57Z","lastTransitionTime":"2025-10-09T10:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.201996 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.202055 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.202066 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.202086 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.202100 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:57Z","lastTransitionTime":"2025-10-09T10:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.305610 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.305672 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.305686 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.305709 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.305723 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:57Z","lastTransitionTime":"2025-10-09T10:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.408657 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.408717 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.408730 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.408769 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.408785 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:57Z","lastTransitionTime":"2025-10-09T10:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.512157 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.512220 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.512240 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.512267 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.512290 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:57Z","lastTransitionTime":"2025-10-09T10:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.601402 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.601465 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.601417 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:57 crc kubenswrapper[4923]: E1009 10:05:57.601639 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:57 crc kubenswrapper[4923]: E1009 10:05:57.601818 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:57 crc kubenswrapper[4923]: E1009 10:05:57.601932 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.615106 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.615161 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.615175 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.615200 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.615219 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:57Z","lastTransitionTime":"2025-10-09T10:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.718253 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.718323 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.718341 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.718368 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.718384 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:57Z","lastTransitionTime":"2025-10-09T10:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.821281 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.821340 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.821361 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.821384 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.821402 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:57Z","lastTransitionTime":"2025-10-09T10:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.924701 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.924819 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.924832 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.924849 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:57 crc kubenswrapper[4923]: I1009 10:05:57.924859 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:57Z","lastTransitionTime":"2025-10-09T10:05:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.027547 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.027631 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.027645 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.027663 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.027675 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.129694 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.129731 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.129741 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.129769 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.129778 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.232519 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.232567 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.232580 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.232596 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.232608 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.337454 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.337521 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.337540 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.337566 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.337580 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.441057 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.441109 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.441118 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.441137 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.441147 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.476994 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:58 crc kubenswrapper[4923]: E1009 10:05:58.477272 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:58 crc kubenswrapper[4923]: E1009 10:05:58.477397 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs podName:0ffdd04f-94e8-4fca-9676-ef87952debcc nodeName:}" failed. No retries permitted until 2025-10-09 10:06:06.477369678 +0000 UTC m=+52.545551434 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs") pod "network-metrics-daemon-dxnsb" (UID: "0ffdd04f-94e8-4fca-9676-ef87952debcc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.546655 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.547252 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.547270 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.547293 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.547304 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.601349 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:05:58 crc kubenswrapper[4923]: E1009 10:05:58.601511 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.649667 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.649705 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.649714 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.649728 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.649737 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.752415 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.752457 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.752471 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.752486 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.752500 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.856797 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.856855 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.856867 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.856887 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.856902 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.959671 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.959712 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.959721 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.959737 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:58 crc kubenswrapper[4923]: I1009 10:05:58.959766 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:58Z","lastTransitionTime":"2025-10-09T10:05:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.062925 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.062991 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.063002 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.063019 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.063028 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.166157 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.166208 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.166218 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.166237 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.166257 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.269604 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.269662 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.269675 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.269701 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.269716 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.373090 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.373165 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.373180 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.373203 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.373218 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.475768 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.475811 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.475823 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.475839 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.475851 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.578860 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.578906 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.578918 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.578940 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.578958 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.600324 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.600374 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:05:59 crc kubenswrapper[4923]: E1009 10:05:59.600494 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.600393 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:05:59 crc kubenswrapper[4923]: E1009 10:05:59.600604 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:05:59 crc kubenswrapper[4923]: E1009 10:05:59.600907 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.681742 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.681825 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.681838 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.681859 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.681876 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.784865 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.784926 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.784940 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.784964 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.785003 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.886777 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.886809 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.886818 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.886832 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.886858 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.989791 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.989855 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.989868 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.989887 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:05:59 crc kubenswrapper[4923]: I1009 10:05:59.989900 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:05:59Z","lastTransitionTime":"2025-10-09T10:05:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.093715 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.093783 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.093798 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.093819 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.093833 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:00Z","lastTransitionTime":"2025-10-09T10:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.199878 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.199952 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.199966 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.199986 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.199999 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:00Z","lastTransitionTime":"2025-10-09T10:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.302536 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.302570 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.302581 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.302593 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.302605 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:00Z","lastTransitionTime":"2025-10-09T10:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.404827 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.404890 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.404905 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.404921 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.404935 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:00Z","lastTransitionTime":"2025-10-09T10:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.506573 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.506611 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.506619 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.506631 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.506640 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:00Z","lastTransitionTime":"2025-10-09T10:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.600864 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:00 crc kubenswrapper[4923]: E1009 10:06:00.601049 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.601827 4923 scope.go:117] "RemoveContainer" containerID="c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.610015 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.610068 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.610081 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.610099 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.610113 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:00Z","lastTransitionTime":"2025-10-09T10:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.622525 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.638785 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.663429 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:47Z\\\",\\\"message\\\":\\\"16 6322 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1009 10:05:47.775227 6322 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-frh4j in node crc\\\\nF1009 10:05:47.775227 6322 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 10:05:47.775228 6322 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1009 10:05:47.775241 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.696291 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.713068 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.713112 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.713127 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.713154 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.713169 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:00Z","lastTransitionTime":"2025-10-09T10:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.721955 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.741641 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.759181 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.784790 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.799099 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.812011 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.815406 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.815439 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.815447 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.815463 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.815473 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:00Z","lastTransitionTime":"2025-10-09T10:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.823883 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.839867 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.853454 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.865638 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.877362 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.888549 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/1.log" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.890832 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:00Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.891977 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244"} Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.918740 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.918827 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.918842 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.918862 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:00 crc kubenswrapper[4923]: I1009 10:06:00.918875 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:00Z","lastTransitionTime":"2025-10-09T10:06:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.021821 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.021897 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.021908 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.021926 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.021937 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.124672 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.124944 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.125118 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.125210 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.125290 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.228426 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.228727 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.228817 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.228885 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.228942 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.331729 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.331793 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.331809 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.331827 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.331840 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.436205 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.436239 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.436249 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.436265 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.436278 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.538922 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.538967 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.538978 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.538994 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.539005 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.600932 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.600974 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.600948 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:01 crc kubenswrapper[4923]: E1009 10:06:01.601068 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:01 crc kubenswrapper[4923]: E1009 10:06:01.601106 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:01 crc kubenswrapper[4923]: E1009 10:06:01.601151 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.641209 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.641250 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.641260 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.641277 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.641288 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.744069 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.744120 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.744130 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.744146 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.744157 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.752741 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.752827 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.752838 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.752864 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.752875 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: E1009 10:06:01.767372 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.770967 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.771010 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.771023 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.771041 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.771052 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: E1009 10:06:01.784700 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.789240 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.789297 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.789310 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.789331 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.789346 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: E1009 10:06:01.804449 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.808489 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.808533 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.808542 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.808557 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.808568 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: E1009 10:06:01.824282 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.829721 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.829767 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.829775 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.829789 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.829799 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: E1009 10:06:01.843323 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: E1009 10:06:01.843452 4923 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.847171 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.847224 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.847238 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.847259 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.847275 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.894906 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.910559 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.920342 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.932962 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.946326 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.949249 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.949281 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.949291 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.949306 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.949316 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:01Z","lastTransitionTime":"2025-10-09T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.959250 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:01 crc kubenswrapper[4923]: I1009 10:06:01.993440 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:01Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.012051 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:47Z\\\",\\\"message\\\":\\\"16 6322 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1009 10:05:47.775227 6322 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-frh4j in node crc\\\\nF1009 10:05:47.775227 6322 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 10:05:47.775228 6322 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1009 10:05:47.775241 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.028934 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.040330 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.052130 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.052171 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.052182 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.052199 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.052210 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:02Z","lastTransitionTime":"2025-10-09T10:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.053554 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.065345 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.078844 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.091603 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.104418 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.116230 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.128848 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.155189 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.155236 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.155247 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.155266 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.155278 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:02Z","lastTransitionTime":"2025-10-09T10:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.258560 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.258614 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.258625 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.258650 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.258663 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:02Z","lastTransitionTime":"2025-10-09T10:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.361742 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.361812 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.361822 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.361837 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.361849 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:02Z","lastTransitionTime":"2025-10-09T10:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.514722 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.514813 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.514829 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.514853 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.514866 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:02Z","lastTransitionTime":"2025-10-09T10:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.601078 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:02 crc kubenswrapper[4923]: E1009 10:06:02.601327 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.617190 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.617233 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.617243 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.617257 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.617266 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:02Z","lastTransitionTime":"2025-10-09T10:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.719306 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.719337 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.719347 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.719360 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.719370 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:02Z","lastTransitionTime":"2025-10-09T10:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.822239 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.822277 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.822288 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.822303 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.822314 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:02Z","lastTransitionTime":"2025-10-09T10:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.901214 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/2.log" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.901988 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/1.log" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.905009 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244" exitCode=1 Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.905065 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.905713 4923 scope.go:117] "RemoveContainer" containerID="c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.906098 4923 scope.go:117] "RemoveContainer" containerID="b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244" Oct 09 10:06:02 crc kubenswrapper[4923]: E1009 10:06:02.906277 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\"" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.924431 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.925878 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.925933 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.925955 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.925982 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.925997 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:02Z","lastTransitionTime":"2025-10-09T10:06:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.943159 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.966139 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2ad8ea5be4258fa1ee8bbd0f4c4b313960edbd22c0f198d2b9ce673d46951a8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:05:47Z\\\",\\\"message\\\":\\\"16 6322 services_controller.go:443] Built service default/kubernetes LB cluster-wide configs for network=default: []services.lbConfig(nil)\\\\nI1009 10:05:47.775227 6322 ovn.go:134] Ensuring zone local for Pod openshift-machine-config-operator/machine-config-daemon-frh4j in node crc\\\\nF1009 10:05:47.775227 6322 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:05:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 10:05:47.775228 6322 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-operator-lifecycle-manager/olm-operator-metrics\\\\\\\"}\\\\nI1009 10:05:47.775241 6\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:47Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.980489 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:02 crc kubenswrapper[4923]: I1009 10:06:02.993556 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:02Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.004011 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.018145 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.028535 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.028777 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.028893 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.028994 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.029087 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.032273 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.045680 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.057913 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.068204 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.080514 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.091293 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.102426 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.112594 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.121132 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.131903 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.131939 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.131950 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.131965 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.131980 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.234044 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.234150 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.234159 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.234173 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.234184 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.336937 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.337251 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.337360 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.337466 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.337568 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.440139 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.440704 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.440830 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.440911 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.440974 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.544110 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.544456 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.544543 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.544620 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.544677 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.600855 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:03 crc kubenswrapper[4923]: E1009 10:06:03.601034 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.600864 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.601096 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:03 crc kubenswrapper[4923]: E1009 10:06:03.601159 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:03 crc kubenswrapper[4923]: E1009 10:06:03.601299 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.647947 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.647984 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.647993 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.648007 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.648018 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.749777 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.749824 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.749833 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.749848 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.749858 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.852448 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.852491 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.852501 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.852516 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.852526 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.909963 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/2.log" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.912798 4923 scope.go:117] "RemoveContainer" containerID="b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244" Oct 09 10:06:03 crc kubenswrapper[4923]: E1009 10:06:03.912973 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\"" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.923958 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.932510 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.941275 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.951388 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.954400 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.954437 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.954447 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.954461 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.954471 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:03Z","lastTransitionTime":"2025-10-09T10:06:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.963727 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.979455 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:03 crc kubenswrapper[4923]: I1009 10:06:03.992223 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:03Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.003469 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.015285 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.028091 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.040558 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.051822 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.056113 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.056142 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.056151 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.056163 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.056172 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.062866 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.074189 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.085322 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.095909 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.158401 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.158447 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.158459 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.158474 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.158485 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.260621 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.260689 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.260699 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.260716 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.260729 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.362858 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.362933 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.362942 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.362956 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.362965 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.465435 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.465487 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.465504 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.465522 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.465534 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.568406 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.568450 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.568461 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.568476 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.568488 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.601276 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:04 crc kubenswrapper[4923]: E1009 10:06:04.601439 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.612935 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.624345 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.639082 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.652711 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.668912 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.670972 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.671020 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.671033 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.671050 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.671063 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.689228 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.702889 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.712714 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.724264 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.736346 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.745860 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.756837 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.769974 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.773994 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.774030 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.774039 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.774052 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.774062 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.795280 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.811026 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.827815 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:04Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.876651 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.876705 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.876716 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.876733 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.876744 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.979195 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.979244 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.979263 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.979281 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:04 crc kubenswrapper[4923]: I1009 10:06:04.979292 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:04Z","lastTransitionTime":"2025-10-09T10:06:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.082472 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.082512 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.082522 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.082538 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.082550 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:05Z","lastTransitionTime":"2025-10-09T10:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.184501 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.184541 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.184550 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.184566 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.184576 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:05Z","lastTransitionTime":"2025-10-09T10:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.286483 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.286590 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.286609 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.286626 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.286636 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:05Z","lastTransitionTime":"2025-10-09T10:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.388789 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.388851 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.388863 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.388880 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.388891 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:05Z","lastTransitionTime":"2025-10-09T10:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.491879 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.491941 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.491954 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.491971 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.491983 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:05Z","lastTransitionTime":"2025-10-09T10:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.595189 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.595238 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.595249 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.595272 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.595291 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:05Z","lastTransitionTime":"2025-10-09T10:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.601146 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:05 crc kubenswrapper[4923]: E1009 10:06:05.601333 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.601447 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.601498 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:05 crc kubenswrapper[4923]: E1009 10:06:05.601597 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:05 crc kubenswrapper[4923]: E1009 10:06:05.601689 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.698085 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.698156 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.698169 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.698192 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.698206 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:05Z","lastTransitionTime":"2025-10-09T10:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.801536 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.801578 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.801589 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.801603 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.801612 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:05Z","lastTransitionTime":"2025-10-09T10:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.904966 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.905011 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.905029 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.905083 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:05 crc kubenswrapper[4923]: I1009 10:06:05.905099 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:05Z","lastTransitionTime":"2025-10-09T10:06:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.008992 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.009037 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.009050 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.009072 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.009088 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.112485 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.112548 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.112558 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.112575 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.112585 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.215796 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.216277 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.216379 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.216496 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.216579 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.319787 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.319839 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.319852 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.319873 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.319885 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.422793 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.422846 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.422863 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.422885 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.422902 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.526050 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.526090 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.526099 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.526136 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.526148 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.567264 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:06 crc kubenswrapper[4923]: E1009 10:06:06.567495 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:06:06 crc kubenswrapper[4923]: E1009 10:06:06.567581 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs podName:0ffdd04f-94e8-4fca-9676-ef87952debcc nodeName:}" failed. No retries permitted until 2025-10-09 10:06:22.5675541 +0000 UTC m=+68.635735856 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs") pod "network-metrics-daemon-dxnsb" (UID: "0ffdd04f-94e8-4fca-9676-ef87952debcc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.601343 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:06 crc kubenswrapper[4923]: E1009 10:06:06.601588 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.629399 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.629455 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.629467 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.629485 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.629497 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.732551 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.732601 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.732619 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.732637 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.732652 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.835698 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.835803 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.835818 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.835838 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.835850 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.939124 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.939191 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.939210 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.939236 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:06 crc kubenswrapper[4923]: I1009 10:06:06.939256 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:06Z","lastTransitionTime":"2025-10-09T10:06:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.042207 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.042262 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.042272 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.042289 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.042300 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.146701 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.146782 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.146794 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.146814 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.146830 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.250074 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.250139 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.250152 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.250175 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.250190 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.352770 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.352818 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.352828 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.352842 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.352852 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.375419 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.375574 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.375613 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.375717 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:06:39.375671267 +0000 UTC m=+85.443853023 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.375780 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.375725 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.375846 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:06:39.375825582 +0000 UTC m=+85.444007338 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.375872 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.375897 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:06:39.375881583 +0000 UTC m=+85.444063339 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.375965 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.375982 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.375995 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.376049 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 10:06:39.376038407 +0000 UTC m=+85.444220353 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.455803 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.455863 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.455873 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.455896 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.455909 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.477324 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.477627 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.477686 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.477702 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.477824 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 10:06:39.477791521 +0000 UTC m=+85.545973477 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.558561 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.558596 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.558619 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.558653 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.558663 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.601008 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.601097 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.601017 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.601534 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.601584 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:07 crc kubenswrapper[4923]: E1009 10:06:07.601613 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.664554 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.664601 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.664611 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.664632 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.664645 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.767152 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.767189 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.767197 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.767210 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.767220 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.869878 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.869921 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.869934 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.869951 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.869963 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.972856 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.972914 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.972924 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.972937 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:07 crc kubenswrapper[4923]: I1009 10:06:07.972945 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:07Z","lastTransitionTime":"2025-10-09T10:06:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.074704 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.074743 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.074776 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.074794 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.074805 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:08Z","lastTransitionTime":"2025-10-09T10:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.178239 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.178293 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.178307 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.178522 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.178535 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:08Z","lastTransitionTime":"2025-10-09T10:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.281501 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.281620 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.281632 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.281650 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.281662 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:08Z","lastTransitionTime":"2025-10-09T10:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.384918 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.384979 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.384992 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.385011 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.385073 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:08Z","lastTransitionTime":"2025-10-09T10:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.487876 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.487930 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.487945 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.487964 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.487974 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:08Z","lastTransitionTime":"2025-10-09T10:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.591042 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.591129 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.591148 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.591181 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.591200 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:08Z","lastTransitionTime":"2025-10-09T10:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.600477 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:08 crc kubenswrapper[4923]: E1009 10:06:08.600662 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.694208 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.694268 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.694280 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.694304 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.694323 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:08Z","lastTransitionTime":"2025-10-09T10:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.796834 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.796892 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.796910 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.796937 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.796954 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:08Z","lastTransitionTime":"2025-10-09T10:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.898864 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.898912 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.898930 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.898948 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:08 crc kubenswrapper[4923]: I1009 10:06:08.898959 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:08Z","lastTransitionTime":"2025-10-09T10:06:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.001501 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.001561 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.001572 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.001588 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.001598 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.097358 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.104265 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.104318 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.104331 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.104355 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.104370 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.112978 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.114644 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.129440 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.151838 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.173307 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.190267 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.207440 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.207534 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.207561 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.207595 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.207619 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.211044 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.227300 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.242467 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.259648 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.271796 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.285820 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.298034 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.310434 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.310467 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.310477 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.310494 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.310507 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.314472 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.329155 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.352615 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.369984 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:09Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.416284 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.416335 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.416348 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.416365 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.416377 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.519161 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.519208 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.519217 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.519234 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.519244 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.601080 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:09 crc kubenswrapper[4923]: E1009 10:06:09.601241 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.601106 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:09 crc kubenswrapper[4923]: E1009 10:06:09.601337 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.601095 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:09 crc kubenswrapper[4923]: E1009 10:06:09.601416 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.621589 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.621644 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.621656 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.621675 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.621693 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.723857 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.723892 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.723901 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.723915 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.723924 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.826244 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.826280 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.826358 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.826376 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.826479 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.928975 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.929019 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.929029 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.929054 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:09 crc kubenswrapper[4923]: I1009 10:06:09.929064 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:09Z","lastTransitionTime":"2025-10-09T10:06:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.030715 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.031078 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.031190 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.031307 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.031393 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.134322 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.134363 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.134373 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.134388 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.134399 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.237263 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.237311 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.237321 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.237336 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.237345 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.340053 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.340100 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.340113 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.340130 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.340142 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.443170 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.443220 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.443234 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.443286 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.443298 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.545769 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.545828 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.545843 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.545865 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.545880 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.601248 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:10 crc kubenswrapper[4923]: E1009 10:06:10.601432 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.648322 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.648382 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.648399 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.648424 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.648444 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.751403 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.751462 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.751476 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.751500 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.751516 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.855031 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.855089 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.855098 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.855118 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.855127 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.956894 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.956931 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.956940 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.956953 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:10 crc kubenswrapper[4923]: I1009 10:06:10.956963 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:10Z","lastTransitionTime":"2025-10-09T10:06:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.059044 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.059082 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.059090 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.059133 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.059142 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.162245 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.162317 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.162325 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.162343 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.162361 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.265673 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.265783 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.265795 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.265813 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.265825 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.369172 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.369271 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.369298 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.369339 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.369364 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.471898 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.471948 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.471961 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.471978 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.471989 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.574432 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.574500 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.574520 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.574551 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.574571 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.601049 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.601186 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:11 crc kubenswrapper[4923]: E1009 10:06:11.601246 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.601062 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:11 crc kubenswrapper[4923]: E1009 10:06:11.601488 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:11 crc kubenswrapper[4923]: E1009 10:06:11.601606 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.678248 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.678307 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.678325 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.678351 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.678370 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.781546 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.781587 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.781597 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.781616 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.781629 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.884511 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.884570 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.884581 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.884600 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.884612 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.987152 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.987201 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.987215 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.987241 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:11 crc kubenswrapper[4923]: I1009 10:06:11.987253 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:11Z","lastTransitionTime":"2025-10-09T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.062196 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.062242 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.062253 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.062269 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.062282 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: E1009 10:06:12.076799 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:12Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.081617 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.081694 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.081708 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.081738 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.081783 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: E1009 10:06:12.096391 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:12Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.099383 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.099443 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.099460 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.099507 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.099522 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: E1009 10:06:12.111736 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:12Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.115151 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.115190 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.115201 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.115217 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.115228 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: E1009 10:06:12.127678 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:12Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.131795 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.131831 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.131841 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.131855 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.131865 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: E1009 10:06:12.146203 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:12Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:12 crc kubenswrapper[4923]: E1009 10:06:12.146511 4923 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.148790 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.148834 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.148872 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.148898 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.148915 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.251656 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.251706 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.251719 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.251743 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.251795 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.355726 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.355862 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.355879 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.355901 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.355918 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.459397 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.459466 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.459476 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.459494 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.459523 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.561354 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.561393 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.561404 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.561419 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.561430 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.600417 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:12 crc kubenswrapper[4923]: E1009 10:06:12.600553 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.665208 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.665242 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.665251 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.665264 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.665296 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.768080 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.768137 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.768148 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.768163 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.768172 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.870898 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.870956 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.870968 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.870986 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.870999 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.974233 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.974282 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.974293 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.974307 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:12 crc kubenswrapper[4923]: I1009 10:06:12.974318 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:12Z","lastTransitionTime":"2025-10-09T10:06:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.077741 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.077812 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.077861 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.077877 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.077887 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:13Z","lastTransitionTime":"2025-10-09T10:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.181530 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.181575 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.181586 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.181610 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.181623 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:13Z","lastTransitionTime":"2025-10-09T10:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.284238 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.284293 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.284304 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.284319 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.284364 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:13Z","lastTransitionTime":"2025-10-09T10:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.386684 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.386734 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.386766 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.386786 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.386797 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:13Z","lastTransitionTime":"2025-10-09T10:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.488906 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.488936 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.488945 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.488958 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.488966 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:13Z","lastTransitionTime":"2025-10-09T10:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.592042 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.592106 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.592115 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.592131 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.592144 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:13Z","lastTransitionTime":"2025-10-09T10:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.600268 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.600276 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:13 crc kubenswrapper[4923]: E1009 10:06:13.600417 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.600276 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:13 crc kubenswrapper[4923]: E1009 10:06:13.600615 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:13 crc kubenswrapper[4923]: E1009 10:06:13.600533 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.694433 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.694483 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.694496 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.694532 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.694544 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:13Z","lastTransitionTime":"2025-10-09T10:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.796669 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.796745 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.796784 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.796800 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.796812 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:13Z","lastTransitionTime":"2025-10-09T10:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.899612 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.899665 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.899678 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.899698 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:13 crc kubenswrapper[4923]: I1009 10:06:13.899711 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:13Z","lastTransitionTime":"2025-10-09T10:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.002282 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.002310 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.002318 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.002330 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.002339 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.104110 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.104146 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.104158 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.104171 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.104180 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.206564 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.206616 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.206629 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.206662 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.206673 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.309027 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.309074 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.309084 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.309099 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.309109 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.411920 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.411964 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.411976 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.411991 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.412000 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.514364 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.514409 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.514420 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.514434 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.514445 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.600306 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:14 crc kubenswrapper[4923]: E1009 10:06:14.600770 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.616787 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.616785 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.616831 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.616999 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.617023 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.617033 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.628549 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.639397 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.653601 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.664186 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.678512 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.697275 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.707222 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.717180 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e6a7e4d-3d70-49c9-af3c-46bbc92e8ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef9abe2735aef05ea4fb2ea269f806a4b4e4cc75d890aedc26b61e00890380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae77d7fd42982b8c1531d352ed15257c3626fe61672015a7ea3a1c7b88db9ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e9c6706f9f71acc1376c0a0def4c117db567dc5b60962106e6d0aebb43caaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.718718 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.718745 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.718768 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.718782 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.718791 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.729207 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.738888 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.750351 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.762296 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.778450 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.791478 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.801037 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.810307 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:14Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.820887 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.820937 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.820947 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.820962 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.820972 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.923065 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.923106 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.923117 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.923132 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:14 crc kubenswrapper[4923]: I1009 10:06:14.923141 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:14Z","lastTransitionTime":"2025-10-09T10:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.025944 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.025976 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.025985 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.025997 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.026005 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.128464 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.128508 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.128516 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.128529 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.128538 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.231377 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.231420 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.231428 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.231443 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.231454 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.334591 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.334630 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.334638 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.334651 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.334661 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.437169 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.437209 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.437219 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.437238 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.437248 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.539741 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.539796 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.539808 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.539825 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.539836 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.600794 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.600876 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.600905 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:15 crc kubenswrapper[4923]: E1009 10:06:15.600952 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:15 crc kubenswrapper[4923]: E1009 10:06:15.601012 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:15 crc kubenswrapper[4923]: E1009 10:06:15.601118 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.642367 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.642417 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.642431 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.642451 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.642464 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.745351 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.745388 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.745396 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.745410 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.745419 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.847814 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.847857 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.847867 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.847881 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.847893 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.949871 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.949928 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.949938 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.949960 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:15 crc kubenswrapper[4923]: I1009 10:06:15.949970 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:15Z","lastTransitionTime":"2025-10-09T10:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.051734 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.051797 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.051806 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.051821 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.051831 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.154561 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.154609 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.154622 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.154643 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.154656 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.257819 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.257879 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.257890 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.257905 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.257916 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.361383 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.361439 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.361463 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.361487 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.361501 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.463602 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.463640 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.463651 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.463670 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.463680 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.566397 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.566461 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.566472 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.566527 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.566550 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.600446 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:16 crc kubenswrapper[4923]: E1009 10:06:16.600618 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.669207 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.669248 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.669260 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.669277 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.669290 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.771728 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.771793 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.771807 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.771822 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.771832 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.873902 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.873932 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.873940 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.873954 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.873965 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.976019 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.976069 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.976080 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.976097 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:16 crc kubenswrapper[4923]: I1009 10:06:16.976111 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:16Z","lastTransitionTime":"2025-10-09T10:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.080116 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.080154 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.080164 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.080178 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.080188 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:17Z","lastTransitionTime":"2025-10-09T10:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.183267 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.183905 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.183950 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.183975 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.183991 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:17Z","lastTransitionTime":"2025-10-09T10:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.286627 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.286674 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.286688 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.286704 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.286717 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:17Z","lastTransitionTime":"2025-10-09T10:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.391118 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.391157 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.391165 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.391178 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.391188 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:17Z","lastTransitionTime":"2025-10-09T10:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.493583 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.493624 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.493634 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.493650 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.493661 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:17Z","lastTransitionTime":"2025-10-09T10:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.597365 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.597424 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.597436 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.597455 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.597465 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:17Z","lastTransitionTime":"2025-10-09T10:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.600744 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.600730 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.600930 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:17 crc kubenswrapper[4923]: E1009 10:06:17.601019 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:17 crc kubenswrapper[4923]: E1009 10:06:17.601102 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:17 crc kubenswrapper[4923]: E1009 10:06:17.601248 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.700484 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.700529 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.700569 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.700587 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.700597 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:17Z","lastTransitionTime":"2025-10-09T10:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.804183 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.804246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.804260 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.804280 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.804294 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:17Z","lastTransitionTime":"2025-10-09T10:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.907397 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.907440 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.907450 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.907463 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:17 crc kubenswrapper[4923]: I1009 10:06:17.907474 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:17Z","lastTransitionTime":"2025-10-09T10:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.010308 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.010345 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.010355 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.010369 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.010378 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.113687 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.113791 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.113860 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.113891 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.113914 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.222339 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.222410 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.222423 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.222446 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.222461 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.325543 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.325591 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.325601 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.325616 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.325626 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.433420 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.433484 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.433495 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.433511 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.433520 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.536665 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.536712 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.536722 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.536739 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.536767 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.600789 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:18 crc kubenswrapper[4923]: E1009 10:06:18.600990 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.639229 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.639306 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.639318 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.639339 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.639354 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.741695 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.741733 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.741744 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.741778 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.741788 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.844088 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.844144 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.844154 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.844170 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.844179 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.946712 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.946777 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.946790 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.946806 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:18 crc kubenswrapper[4923]: I1009 10:06:18.946816 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:18Z","lastTransitionTime":"2025-10-09T10:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.050417 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.050466 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.050476 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.050494 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.050506 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.152357 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.152396 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.152405 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.152418 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.152427 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.255578 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.255670 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.255681 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.255709 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.255723 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.358022 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.358071 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.358084 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.358104 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.358114 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.461313 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.461372 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.461386 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.461411 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.461426 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.564084 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.564134 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.564145 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.564197 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.564213 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.600826 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.600846 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.600926 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:19 crc kubenswrapper[4923]: E1009 10:06:19.600969 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:19 crc kubenswrapper[4923]: E1009 10:06:19.601086 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:19 crc kubenswrapper[4923]: E1009 10:06:19.601431 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.601980 4923 scope.go:117] "RemoveContainer" containerID="b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244" Oct 09 10:06:19 crc kubenswrapper[4923]: E1009 10:06:19.602205 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\"" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.666483 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.666529 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.666540 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.666557 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.666569 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.769433 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.769478 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.769489 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.769503 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.769513 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.872126 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.872166 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.872177 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.872195 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.872206 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.974207 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.974248 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.974259 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.974273 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:19 crc kubenswrapper[4923]: I1009 10:06:19.974284 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:19Z","lastTransitionTime":"2025-10-09T10:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.076404 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.076441 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.076452 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.076471 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.076483 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:20Z","lastTransitionTime":"2025-10-09T10:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.178593 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.178653 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.178662 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.178677 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.178697 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:20Z","lastTransitionTime":"2025-10-09T10:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.280734 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.280804 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.280813 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.280833 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.280843 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:20Z","lastTransitionTime":"2025-10-09T10:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.382763 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.382804 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.382812 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.382824 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.382833 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:20Z","lastTransitionTime":"2025-10-09T10:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.485676 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.485716 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.485731 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.485746 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.485783 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:20Z","lastTransitionTime":"2025-10-09T10:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.588431 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.588468 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.588478 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.588493 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.588504 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:20Z","lastTransitionTime":"2025-10-09T10:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.600916 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:20 crc kubenswrapper[4923]: E1009 10:06:20.601050 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.691392 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.691427 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.691436 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.691449 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.691458 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:20Z","lastTransitionTime":"2025-10-09T10:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.794890 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.794991 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.795009 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.795035 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.795047 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:20Z","lastTransitionTime":"2025-10-09T10:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.898337 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.898392 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.898403 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.898420 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:20 crc kubenswrapper[4923]: I1009 10:06:20.898774 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:20Z","lastTransitionTime":"2025-10-09T10:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.000905 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.000939 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.000947 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.000960 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.000970 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.103320 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.103376 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.103390 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.103407 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.103421 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.207285 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.207332 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.207343 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.207362 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.207373 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.309290 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.309331 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.309340 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.309353 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.309365 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.411284 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.411322 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.411332 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.411347 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.411357 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.514359 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.514402 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.514412 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.514427 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.514438 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.600462 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.600570 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:21 crc kubenswrapper[4923]: E1009 10:06:21.600631 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:21 crc kubenswrapper[4923]: E1009 10:06:21.600662 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.600689 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:21 crc kubenswrapper[4923]: E1009 10:06:21.600779 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.616122 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.616162 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.616175 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.616193 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.616202 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.718802 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.718837 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.718846 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.718859 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.718868 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.821185 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.821469 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.821569 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.821659 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.821730 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.924072 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.924123 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.924134 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.924147 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:21 crc kubenswrapper[4923]: I1009 10:06:21.924156 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:21Z","lastTransitionTime":"2025-10-09T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.026706 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.026767 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.026777 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.026792 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.026801 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.129291 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.129336 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.129347 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.129361 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.129370 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.231352 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.231401 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.231412 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.231430 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.231442 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.335381 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.335439 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.335453 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.335474 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.335487 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.438882 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.438950 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.438964 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.438989 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.439006 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.542224 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.542300 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.542313 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.542336 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.542350 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.543425 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.543499 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.543517 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.543542 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.543557 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: E1009 10:06:22.559424 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:22Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.563505 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.563568 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.563584 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.563604 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.563621 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: E1009 10:06:22.578895 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:22Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.583504 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.583667 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.583774 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.583900 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.583983 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.601063 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:22 crc kubenswrapper[4923]: E1009 10:06:22.601441 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:22 crc kubenswrapper[4923]: E1009 10:06:22.601807 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:22Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.612645 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.612785 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.612907 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.613147 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.613488 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: E1009 10:06:22.631407 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:22Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.637559 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.637606 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.637617 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.637640 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.637652 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.640665 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:22 crc kubenswrapper[4923]: E1009 10:06:22.640836 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:06:22 crc kubenswrapper[4923]: E1009 10:06:22.640900 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs podName:0ffdd04f-94e8-4fca-9676-ef87952debcc nodeName:}" failed. No retries permitted until 2025-10-09 10:06:54.640876215 +0000 UTC m=+100.709057971 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs") pod "network-metrics-daemon-dxnsb" (UID: "0ffdd04f-94e8-4fca-9676-ef87952debcc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:06:22 crc kubenswrapper[4923]: E1009 10:06:22.651408 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:22Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:22Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:22 crc kubenswrapper[4923]: E1009 10:06:22.651585 4923 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.654292 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.654345 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.654354 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.654411 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.654425 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.756520 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.756567 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.756579 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.756595 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.756608 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.858732 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.858809 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.858824 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.858842 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.858854 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.961398 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.961432 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.961451 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.961467 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:22 crc kubenswrapper[4923]: I1009 10:06:22.961479 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:22Z","lastTransitionTime":"2025-10-09T10:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.064144 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.064187 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.064198 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.064213 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.064226 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.166975 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.167014 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.167025 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.167043 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.167055 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.270659 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.270723 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.270741 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.270792 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.270812 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.373664 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.373720 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.373735 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.373783 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.373819 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.476126 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.476435 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.476504 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.476567 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.476628 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.579834 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.579870 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.579884 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.579907 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.579919 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.600396 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.600422 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:23 crc kubenswrapper[4923]: E1009 10:06:23.600540 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:23 crc kubenswrapper[4923]: E1009 10:06:23.600634 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.600650 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:23 crc kubenswrapper[4923]: E1009 10:06:23.600903 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.682746 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.682837 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.682852 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.682875 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.682889 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.786053 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.786115 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.786130 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.786152 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.786170 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.889685 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.889733 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.889743 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.889782 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.889794 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.992568 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.992626 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.992640 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.992661 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:23 crc kubenswrapper[4923]: I1009 10:06:23.992677 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:23Z","lastTransitionTime":"2025-10-09T10:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.095441 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.095474 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.095482 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.095497 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.095508 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:24Z","lastTransitionTime":"2025-10-09T10:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.197248 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.197277 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.197285 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.197299 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.197309 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:24Z","lastTransitionTime":"2025-10-09T10:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.299898 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.299959 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.299975 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.299995 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.300007 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:24Z","lastTransitionTime":"2025-10-09T10:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.401931 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.401963 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.401973 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.401988 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.402003 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:24Z","lastTransitionTime":"2025-10-09T10:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.516569 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.516618 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.516628 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.516643 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.516655 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:24Z","lastTransitionTime":"2025-10-09T10:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.601282 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:24 crc kubenswrapper[4923]: E1009 10:06:24.601404 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.616148 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.623586 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.623675 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.623686 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.623701 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.623711 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:24Z","lastTransitionTime":"2025-10-09T10:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.634006 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.647271 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.660788 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e6a7e4d-3d70-49c9-af3c-46bbc92e8ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef9abe2735aef05ea4fb2ea269f806a4b4e4cc75d890aedc26b61e00890380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae77d7fd42982b8c1531d352ed15257c3626fe61672015a7ea3a1c7b88db9ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e9c6706f9f71acc1376c0a0def4c117db567dc5b60962106e6d0aebb43caaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.674321 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.685448 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.700806 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.716878 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.726460 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.726550 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.726563 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.726579 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.726590 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:24Z","lastTransitionTime":"2025-10-09T10:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.732241 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.752614 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.768837 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.781412 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.795855 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.809392 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.824672 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.829078 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.829118 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.829132 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.829149 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.829163 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:24Z","lastTransitionTime":"2025-10-09T10:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.840275 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.855320 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:24Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.931463 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.931497 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.931508 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.931522 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.931531 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:24Z","lastTransitionTime":"2025-10-09T10:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.977238 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/0.log" Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.977291 4923 generic.go:334] "Generic (PLEG): container finished" podID="d5c5f455-4e94-4f9a-b6fa-2535685c6354" containerID="c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24" exitCode=1 Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.977325 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4cn4f" event={"ID":"d5c5f455-4e94-4f9a-b6fa-2535685c6354","Type":"ContainerDied","Data":"c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24"} Oct 09 10:06:24 crc kubenswrapper[4923]: I1009 10:06:24.977707 4923 scope.go:117] "RemoveContainer" containerID="c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.005257 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.017671 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.028502 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e6a7e4d-3d70-49c9-af3c-46bbc92e8ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef9abe2735aef05ea4fb2ea269f806a4b4e4cc75d890aedc26b61e00890380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae77d7fd42982b8c1531d352ed15257c3626fe61672015a7ea3a1c7b88db9ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e9c6706f9f71acc1376c0a0def4c117db567dc5b60962106e6d0aebb43caaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.034640 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.034708 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.034720 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.034734 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.034746 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.047653 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.062928 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.076275 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.086158 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.096747 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.107806 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.122539 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.135929 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.136885 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.136927 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.136941 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.136957 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.136968 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.151732 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:24Z\\\",\\\"message\\\":\\\"2025-10-09T10:05:38+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc\\\\n2025-10-09T10:05:38+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc to /host/opt/cni/bin/\\\\n2025-10-09T10:05:39Z [verbose] multus-daemon started\\\\n2025-10-09T10:05:39Z [verbose] Readiness Indicator file check\\\\n2025-10-09T10:06:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.169857 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.184367 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.196306 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.209433 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.222334 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.239280 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.239330 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.239343 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.239359 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.239416 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.342317 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.342393 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.342405 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.342425 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.342440 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.445681 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.445738 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.445802 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.445833 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.445849 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.548855 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.548911 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.548924 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.548947 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.548963 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.600660 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.600770 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:25 crc kubenswrapper[4923]: E1009 10:06:25.601206 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:25 crc kubenswrapper[4923]: E1009 10:06:25.601016 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.600874 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:25 crc kubenswrapper[4923]: E1009 10:06:25.601319 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.652071 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.652138 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.652154 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.652179 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.652196 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.754744 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.755081 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.755183 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.755270 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.755341 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.858314 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.858641 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.858715 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.858860 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.858927 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.961787 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.962085 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.962191 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.962271 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.962338 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:25Z","lastTransitionTime":"2025-10-09T10:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.982235 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/0.log" Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.982297 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4cn4f" event={"ID":"d5c5f455-4e94-4f9a-b6fa-2535685c6354","Type":"ContainerStarted","Data":"29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a"} Oct 09 10:06:25 crc kubenswrapper[4923]: I1009 10:06:25.995488 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:25Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.009154 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:24Z\\\",\\\"message\\\":\\\"2025-10-09T10:05:38+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc\\\\n2025-10-09T10:05:38+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc to /host/opt/cni/bin/\\\\n2025-10-09T10:05:39Z [verbose] multus-daemon started\\\\n2025-10-09T10:05:39Z [verbose] Readiness Indicator file check\\\\n2025-10-09T10:06:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.027585 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.040010 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.051986 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.066791 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.066860 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.066871 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.066892 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.066910 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.069065 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.082203 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.095650 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.107341 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.119216 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e6a7e4d-3d70-49c9-af3c-46bbc92e8ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef9abe2735aef05ea4fb2ea269f806a4b4e4cc75d890aedc26b61e00890380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae77d7fd42982b8c1531d352ed15257c3626fe61672015a7ea3a1c7b88db9ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e9c6706f9f71acc1376c0a0def4c117db567dc5b60962106e6d0aebb43caaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.135604 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.152996 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.166206 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.168950 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.168981 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.168990 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.169008 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.169023 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.180028 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.196421 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.211590 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.233485 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:26Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.272603 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.272656 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.272666 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.272681 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.272690 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.375246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.375282 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.375293 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.375308 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.375351 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.477843 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.477916 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.477932 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.477968 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.477983 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.580204 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.580239 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.580249 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.580263 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.580274 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.601242 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:26 crc kubenswrapper[4923]: E1009 10:06:26.601377 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.682464 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.682694 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.682790 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.682875 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.682978 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.785527 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.785831 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.785919 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.785985 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.786045 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.888148 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.888179 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.888188 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.888201 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.888210 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.990288 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.990561 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.990659 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.990776 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:26 crc kubenswrapper[4923]: I1009 10:06:26.990840 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:26Z","lastTransitionTime":"2025-10-09T10:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.093708 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.093770 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.093783 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.093797 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.093806 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:27Z","lastTransitionTime":"2025-10-09T10:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.195904 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.196211 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.196308 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.196410 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.196601 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:27Z","lastTransitionTime":"2025-10-09T10:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.298921 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.298968 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.298980 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.298998 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.299011 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:27Z","lastTransitionTime":"2025-10-09T10:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.402201 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.402245 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.402259 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.402274 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.402284 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:27Z","lastTransitionTime":"2025-10-09T10:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.505336 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.505405 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.505414 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.505427 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.505437 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:27Z","lastTransitionTime":"2025-10-09T10:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.600528 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:27 crc kubenswrapper[4923]: E1009 10:06:27.600679 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.600779 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:27 crc kubenswrapper[4923]: E1009 10:06:27.600838 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.600891 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:27 crc kubenswrapper[4923]: E1009 10:06:27.600948 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.608170 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.608216 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.608227 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.608246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.608259 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:27Z","lastTransitionTime":"2025-10-09T10:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.710143 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.710458 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.710551 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.710641 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.710790 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:27Z","lastTransitionTime":"2025-10-09T10:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.813384 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.813425 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.813434 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.813450 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.813460 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:27Z","lastTransitionTime":"2025-10-09T10:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.915634 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.915951 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.916073 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.916174 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:27 crc kubenswrapper[4923]: I1009 10:06:27.916289 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:27Z","lastTransitionTime":"2025-10-09T10:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.018405 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.018434 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.018444 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.018457 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.018466 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.125293 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.125321 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.125330 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.125353 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.125362 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.227855 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.227896 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.227910 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.227926 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.227939 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.329956 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.329988 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.329996 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.330008 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.330017 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.432059 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.432103 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.432116 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.432130 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.432141 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.534424 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.534469 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.534481 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.534497 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.534511 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.600608 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:28 crc kubenswrapper[4923]: E1009 10:06:28.600805 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.636502 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.636544 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.636555 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.636601 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.636617 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.739540 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.739580 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.739591 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.739610 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.739622 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.842527 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.842580 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.842607 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.842627 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.842640 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.945105 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.945161 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.945171 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.945187 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:28 crc kubenswrapper[4923]: I1009 10:06:28.945200 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:28Z","lastTransitionTime":"2025-10-09T10:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.048047 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.048082 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.048091 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.048104 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.048116 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.150260 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.150306 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.150318 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.150338 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.150350 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.252664 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.252703 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.252711 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.252723 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.252732 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.355062 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.355114 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.355127 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.355144 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.355155 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.457519 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.457546 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.457553 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.457566 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.457576 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.560292 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.560327 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.560337 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.560354 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.560365 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.600520 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:29 crc kubenswrapper[4923]: E1009 10:06:29.600652 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.600878 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:29 crc kubenswrapper[4923]: E1009 10:06:29.600941 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.601115 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:29 crc kubenswrapper[4923]: E1009 10:06:29.601180 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.662909 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.662976 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.662990 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.663005 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.663016 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.766827 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.766871 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.766881 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.766903 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.766915 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.870513 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.870556 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.870565 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.870583 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.870594 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.974554 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.974599 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.974611 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.974629 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:29 crc kubenswrapper[4923]: I1009 10:06:29.974640 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:29Z","lastTransitionTime":"2025-10-09T10:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.078172 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.078436 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.078450 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.078464 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.078481 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:30Z","lastTransitionTime":"2025-10-09T10:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.181035 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.181094 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.181107 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.181127 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.181140 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:30Z","lastTransitionTime":"2025-10-09T10:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.285317 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.285471 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.285489 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.285517 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.285535 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:30Z","lastTransitionTime":"2025-10-09T10:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.389579 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.389642 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.389685 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.389719 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.389742 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:30Z","lastTransitionTime":"2025-10-09T10:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.493444 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.493497 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.493507 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.493525 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.493535 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:30Z","lastTransitionTime":"2025-10-09T10:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.596240 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.596297 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.596309 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.596329 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.596341 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:30Z","lastTransitionTime":"2025-10-09T10:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.600847 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:30 crc kubenswrapper[4923]: E1009 10:06:30.601093 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.603505 4923 scope.go:117] "RemoveContainer" containerID="b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.700000 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.700054 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.700065 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.700084 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.700095 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:30Z","lastTransitionTime":"2025-10-09T10:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.804089 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.804134 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.804147 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.804167 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.804179 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:30Z","lastTransitionTime":"2025-10-09T10:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.906941 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.907031 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.907045 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.907062 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:30 crc kubenswrapper[4923]: I1009 10:06:30.907074 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:30Z","lastTransitionTime":"2025-10-09T10:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.009113 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.009168 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.009179 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.009196 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.009208 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.112144 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.112183 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.112192 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.112207 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.112217 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.214351 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.214391 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.214399 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.214412 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.214421 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.316847 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.316884 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.316893 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.316908 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.316917 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.419410 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.419450 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.419460 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.419478 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.419489 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.522202 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.522263 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.522277 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.522300 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.522313 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.601328 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:31 crc kubenswrapper[4923]: E1009 10:06:31.601454 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.601607 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:31 crc kubenswrapper[4923]: E1009 10:06:31.601650 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.601735 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:31 crc kubenswrapper[4923]: E1009 10:06:31.601812 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.627663 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.627743 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.627774 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.627801 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.627824 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.731127 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.731175 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.731185 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.731202 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.731213 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.834819 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.834883 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.834896 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.834916 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.834931 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.937635 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.937695 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.937706 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.937727 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:31 crc kubenswrapper[4923]: I1009 10:06:31.937738 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:31Z","lastTransitionTime":"2025-10-09T10:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.002054 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/3.log" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.002937 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/2.log" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.006415 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" exitCode=1 Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.006469 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.006521 4923 scope.go:117] "RemoveContainer" containerID="b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.007301 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:06:32 crc kubenswrapper[4923]: E1009 10:06:32.007490 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\"" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.019311 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e6a7e4d-3d70-49c9-af3c-46bbc92e8ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef9abe2735aef05ea4fb2ea269f806a4b4e4cc75d890aedc26b61e00890380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae77d7fd42982b8c1531d352ed15257c3626fe61672015a7ea3a1c7b88db9ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e9c6706f9f71acc1376c0a0def4c117db567dc5b60962106e6d0aebb43caaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.032544 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.040349 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.040405 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.040416 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.040431 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.040443 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.045991 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.060244 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.073867 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.094210 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:31Z\\\",\\\"message\\\":\\\"event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1009 10:06:31.921644 6918 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-nrck8\\\\nI1009 10:06:31.922088 6918 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-nrck8\\\\nI1009 10:06:31.922095 6918 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-nrck8 in node crc\\\\nI1009 10:06:31.922060 6918 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1009 10:06:31.922129 6918 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1009 10:06:31.922100 6918 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-nrck8 after 0 failed attempt(s)\\\\nI1009 10:06:31.922168 6918 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-nrck8\\\\nI1009 10:06:31.921488 6918 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.108373 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.119740 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.131586 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.142988 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.143040 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.143053 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.143073 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.143087 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.146526 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.158888 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.171270 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.188681 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:24Z\\\",\\\"message\\\":\\\"2025-10-09T10:05:38+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc\\\\n2025-10-09T10:05:38+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc to /host/opt/cni/bin/\\\\n2025-10-09T10:05:39Z [verbose] multus-daemon started\\\\n2025-10-09T10:05:39Z [verbose] Readiness Indicator file check\\\\n2025-10-09T10:06:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.203228 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.220176 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.235311 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.245349 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.245398 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.245411 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.245431 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.245447 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.247712 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.349036 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.349097 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.349115 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.349140 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.349157 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.452698 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.452838 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.452880 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.452913 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.452942 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.556354 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.556396 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.556406 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.556421 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.556433 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.601430 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:32 crc kubenswrapper[4923]: E1009 10:06:32.601745 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.660086 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.660195 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.660235 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.660275 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.660302 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.763259 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.763314 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.763330 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.763352 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.763367 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.866577 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.866615 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.866625 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.866638 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.866647 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.887304 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.887359 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.887368 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.887385 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.887401 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: E1009 10:06:32.901162 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.907055 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.907113 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.907131 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.907159 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.907180 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: E1009 10:06:32.929217 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.934569 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.934630 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.934646 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.934677 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.934702 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: E1009 10:06:32.946735 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.950481 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.950523 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.950539 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.950562 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.950579 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: E1009 10:06:32.969711 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.974340 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.974391 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.974401 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.974419 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.974434 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:32 crc kubenswrapper[4923]: E1009 10:06:32.987900 4923 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e7a319c8-81e7-4f84-9486-6dc3c791c3fc\\\",\\\"systemUUID\\\":\\\"4c539cb4-9caf-4858-827b-2e25c2e1065c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:32Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:32 crc kubenswrapper[4923]: E1009 10:06:32.988021 4923 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.990152 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.990188 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.990202 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.990254 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:32 crc kubenswrapper[4923]: I1009 10:06:32.990269 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:32Z","lastTransitionTime":"2025-10-09T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.012730 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/3.log" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.094004 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.094051 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.094060 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.094076 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.094087 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:33Z","lastTransitionTime":"2025-10-09T10:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.197272 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.197366 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.197386 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.197419 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.197439 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:33Z","lastTransitionTime":"2025-10-09T10:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.300154 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.300198 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.300207 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.300222 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.300233 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:33Z","lastTransitionTime":"2025-10-09T10:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.402805 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.402860 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.402873 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.402894 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.402908 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:33Z","lastTransitionTime":"2025-10-09T10:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.506310 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.506381 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.506396 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.506426 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.506443 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:33Z","lastTransitionTime":"2025-10-09T10:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.600969 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.600969 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.601121 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:33 crc kubenswrapper[4923]: E1009 10:06:33.601337 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:33 crc kubenswrapper[4923]: E1009 10:06:33.601449 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:33 crc kubenswrapper[4923]: E1009 10:06:33.601540 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.609011 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.609047 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.609057 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.609070 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.609080 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:33Z","lastTransitionTime":"2025-10-09T10:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.712495 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.712560 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.712585 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.712614 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.712635 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:33Z","lastTransitionTime":"2025-10-09T10:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.815547 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.815590 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.815602 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.815619 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.815630 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:33Z","lastTransitionTime":"2025-10-09T10:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.918341 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.918375 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.918385 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.918398 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:33 crc kubenswrapper[4923]: I1009 10:06:33.918407 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:33Z","lastTransitionTime":"2025-10-09T10:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.020904 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.020949 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.020958 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.020973 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.020984 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.122634 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.122679 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.122687 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.122703 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.122712 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.226428 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.226515 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.226535 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.226559 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.226577 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.329469 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.329535 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.329549 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.329568 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.329580 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.433162 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.433232 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.433249 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.433273 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.433286 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.541287 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.541352 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.541369 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.541390 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.541403 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.601318 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:34 crc kubenswrapper[4923]: E1009 10:06:34.601624 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.618074 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.634307 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.644003 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.644066 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.644080 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.644100 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.644113 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.659633 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b71e2117b1b01e12dc3bbdbf4452bbb1e0b15f3f40093aee2b4810007b6cf244\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:02Z\\\",\\\"message\\\":\\\"001, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1009 10:06:01.678338 6539 obj_retry.go:303] Retry object setup: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678348 6539 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g\\\\nI1009 10:06:01.678358 6539 ovn.go:134] Ensuring zone local for Pod openshift-network-console/networking-console-plugin-85b44fc459-gdk6g in node crc\\\\nI1009 10:06:01.678364 6539 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-target-xd92c] creating logical port openshift-network-diagnostics_network-check-target-xd92c for pod on switch crc\\\\nI1009 10:06:01.678201 6539 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1009 10:06:01.678198 6539 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer bec\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:31Z\\\",\\\"message\\\":\\\"event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1009 10:06:31.921644 6918 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-nrck8\\\\nI1009 10:06:31.922088 6918 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-nrck8\\\\nI1009 10:06:31.922095 6918 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-nrck8 in node crc\\\\nI1009 10:06:31.922060 6918 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1009 10:06:31.922129 6918 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1009 10:06:31.922100 6918 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-nrck8 after 0 failed attempt(s)\\\\nI1009 10:06:31.922168 6918 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-nrck8\\\\nI1009 10:06:31.921488 6918 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.677608 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.696093 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.710142 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.724563 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.741630 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.746111 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.746150 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.746164 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.746186 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.746201 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.756686 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.770927 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:24Z\\\",\\\"message\\\":\\\"2025-10-09T10:05:38+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc\\\\n2025-10-09T10:05:38+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc to /host/opt/cni/bin/\\\\n2025-10-09T10:05:39Z [verbose] multus-daemon started\\\\n2025-10-09T10:05:39Z [verbose] Readiness Indicator file check\\\\n2025-10-09T10:06:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.783904 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.800784 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.815478 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.828603 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.843696 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e6a7e4d-3d70-49c9-af3c-46bbc92e8ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef9abe2735aef05ea4fb2ea269f806a4b4e4cc75d890aedc26b61e00890380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae77d7fd42982b8c1531d352ed15257c3626fe61672015a7ea3a1c7b88db9ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e9c6706f9f71acc1376c0a0def4c117db567dc5b60962106e6d0aebb43caaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.848843 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.848884 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.848895 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.848912 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.848925 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.862205 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.877001 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:34Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.951592 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.951641 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.951653 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.951669 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:34 crc kubenswrapper[4923]: I1009 10:06:34.951680 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:34Z","lastTransitionTime":"2025-10-09T10:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.055201 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.055248 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.055262 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.055279 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.055289 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.159035 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.159076 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.159084 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.159108 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.159118 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.263143 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.263215 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.263231 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.263256 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.263290 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.368098 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.368197 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.368211 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.368232 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.368249 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.471452 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.471503 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.471514 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.471528 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.471539 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.574423 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.574493 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.574506 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.574530 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.574545 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.600875 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:35 crc kubenswrapper[4923]: E1009 10:06:35.601038 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.601105 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.601194 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:35 crc kubenswrapper[4923]: E1009 10:06:35.601300 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:35 crc kubenswrapper[4923]: E1009 10:06:35.601435 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.678883 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.678941 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.678952 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.678974 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.679008 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.782361 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.782425 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.782434 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.782446 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.782454 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.885003 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.885084 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.885094 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.885130 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.885150 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.988018 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.988076 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.988086 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.988105 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:35 crc kubenswrapper[4923]: I1009 10:06:35.988118 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:35Z","lastTransitionTime":"2025-10-09T10:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.091329 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.091583 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.091596 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.091618 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.091631 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:36Z","lastTransitionTime":"2025-10-09T10:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.194420 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.194474 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.194487 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.194507 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.194519 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:36Z","lastTransitionTime":"2025-10-09T10:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.298258 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.298330 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.298346 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.298366 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.298378 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:36Z","lastTransitionTime":"2025-10-09T10:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.401288 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.401365 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.401378 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.401400 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.401413 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:36Z","lastTransitionTime":"2025-10-09T10:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.506191 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.506257 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.506268 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.506287 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.506300 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:36Z","lastTransitionTime":"2025-10-09T10:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.601085 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:36 crc kubenswrapper[4923]: E1009 10:06:36.601277 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.608042 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.608079 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.608088 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.608104 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.608114 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:36Z","lastTransitionTime":"2025-10-09T10:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.711283 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.711344 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.711355 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.711378 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.711393 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:36Z","lastTransitionTime":"2025-10-09T10:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.814264 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.814309 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.814319 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.814338 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.814348 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:36Z","lastTransitionTime":"2025-10-09T10:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.918040 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.918090 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.918101 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.918119 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:36 crc kubenswrapper[4923]: I1009 10:06:36.918131 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:36Z","lastTransitionTime":"2025-10-09T10:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.021235 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.021315 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.021339 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.021372 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.021395 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.124515 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.124565 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.124577 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.124593 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.124605 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.228451 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.228525 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.228538 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.228560 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.228572 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.332125 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.332179 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.332195 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.332220 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.332233 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.435661 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.435696 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.435706 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.435722 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.435733 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.537496 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.537540 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.537552 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.537571 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.537582 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.601236 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.601329 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:37 crc kubenswrapper[4923]: E1009 10:06:37.601459 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.601263 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:37 crc kubenswrapper[4923]: E1009 10:06:37.601612 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:37 crc kubenswrapper[4923]: E1009 10:06:37.601719 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.640464 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.640512 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.640523 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.640543 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.640554 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.743194 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.743299 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.743320 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.743353 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.743373 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.846845 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.846899 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.846915 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.846934 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.846947 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.951618 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.951688 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.951705 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.951732 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:37 crc kubenswrapper[4923]: I1009 10:06:37.951787 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:37Z","lastTransitionTime":"2025-10-09T10:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.055838 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.056325 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.056397 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.056467 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.056546 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.160293 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.160340 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.160349 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.160367 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.160377 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.263066 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.263107 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.263117 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.263131 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.263142 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.269360 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.270528 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:06:38 crc kubenswrapper[4923]: E1009 10:06:38.270863 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\"" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.286662 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"23a187d1-c256-4740-865b-bb4ef267cafb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5f15d976a44ca492afb04b545f17a5c800b9c4d68c6d875f6f19c698cfc1ec0d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d06d5ca3a04d5db7a12f66c44e5530120ec08106e2ad9b5511fd8b04561218fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e0c83a1b67839471df3b99b158efde9376d47277434630d9d3b4bb6014f37f5f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://22faf3db573d1064db47c1d8cb7dd16d615c78355a8fb4babf372e0d65816410\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.303973 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0a266f15572eb750497f3dbb8d225c1cc0b91178a43a24dbbc7b3f98d84576e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://461c034eb66e882b9224125fecf3c9449282220ef0f469334277ab8dfe98f1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.323666 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fb57bacd5111198d4389835f02b5158e0f3256e428bb3f9d64737fef745e6c6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.340951 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-4cn4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5c5f455-4e94-4f9a-b6fa-2535685c6354\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:24Z\\\",\\\"message\\\":\\\"2025-10-09T10:05:38+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc\\\\n2025-10-09T10:05:38+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_c2bc8710-076b-4a72-9cb2-7c32f81619dc to /host/opt/cni/bin/\\\\n2025-10-09T10:05:39Z [verbose] multus-daemon started\\\\n2025-10-09T10:05:39Z [verbose] Readiness Indicator file check\\\\n2025-10-09T10:06:24Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:06:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-khbl2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-4cn4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.357476 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ea1bca1f-8564-4e84-b48e-b105ed075b44\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e2371a54e0a4447df17f050155c7a920dbaf9499b97feed0b2890b5a8c20afc4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://60776d8eed604d91e7937bf700b6cae8245527091e91210b75979b4ce7ffc732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9r9m4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:48Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bxn56\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.365971 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.366023 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.366037 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.366055 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.366066 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.377611 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47c6dc35-0cf1-4b74-86a9-6e8b58d240e2\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://70dd9acef9d6520f39da7fa3864a96cedaa1e3b14e035b9bbd65be546d3f4be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e99067ee30256d10b38bc5ed096d95f160ef5d007b1423e774915cadf67ce205\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4df1b173a051779c32a04c88c1dbad797098ac2488217373c9851f4ffafde655\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ca7e69c7086b7000f0a148b95e63197f3a5887be0dc276357ca6080a3c544ec2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0deaa4749f1ab8ea76441118cee4a3ccfb2cb86b1d6eeec0c90472d0383b26f4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T10:05:34Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 10:05:19.009972 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 10:05:19.011657 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4178364279/tls.crt::/tmp/serving-cert-4178364279/tls.key\\\\\\\"\\\\nI1009 10:05:34.886334 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 10:05:34.892674 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 10:05:34.892729 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 10:05:34.892891 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 10:05:34.892912 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 10:05:34.923107 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 10:05:34.923143 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923148 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 10:05:34.923153 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 10:05:34.923157 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 10:05:34.923160 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 10:05:34.923164 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 10:05:34.923462 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 10:05:34.928459 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:18Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93d3f1f1a09da55ba5d47255b45446873119824436dae0bbb5d86722585cea74\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a181046ebcfd4947e67f619aa4783ec50a5ab392f82c60b6e75816e3e92133\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.391104 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.403079 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-k72w6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"58586169-cd8e-4003-a38e-79c600a6a845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d33d2fc458a357de08f8cf1330aaca483a5241d7515fd2ee19e2b5ab4c7c4bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-szxlf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:39Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-k72w6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.415925 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0e6a7e4d-3d70-49c9-af3c-46bbc92e8ee0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8eef9abe2735aef05ea4fb2ea269f806a4b4e4cc75d890aedc26b61e00890380\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae77d7fd42982b8c1531d352ed15257c3626fe61672015a7ea3a1c7b88db9ae0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e9c6706f9f71acc1376c0a0def4c117db567dc5b60962106e6d0aebb43caaa6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://009c5a8b37189ead5cc8a868f1d7c0aeb04eba81c5f7783bd760dfcb43870572\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:15Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:14Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.432991 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.446030 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-nrck8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7fd8165c-555f-4ded-8081-cb273d195245\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://11568237d62f6e5d405d7c58e9871890f4614a7ff7d442a68afa7ef2518b41ee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xs4hr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-nrck8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.460059 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:35Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.470158 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.470190 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.470199 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.470214 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.470224 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.474492 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9d13bafc1ddcef50e984b3f3fec9e38996e42d3131788b5fb4252ec4ec7426c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.491209 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bc2a175c-508c-4f50-8205-013a9482c989\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T10:06:31Z\\\",\\\"message\\\":\\\"event on pod openshift-network-node-identity/network-node-identity-vrzqb\\\\nI1009 10:06:31.921644 6918 obj_retry.go:303] Retry object setup: *v1.Pod openshift-dns/node-resolver-nrck8\\\\nI1009 10:06:31.922088 6918 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-nrck8\\\\nI1009 10:06:31.922095 6918 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-nrck8 in node crc\\\\nI1009 10:06:31.922060 6918 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1009 10:06:31.922129 6918 obj_retry.go:365] Adding new object: *v1.Pod openshift-kube-apiserver/kube-apiserver-crc\\\\nI1009 10:06:31.922100 6918 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-nrck8 after 0 failed attempt(s)\\\\nI1009 10:06:31.922168 6918 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-nrck8\\\\nI1009 10:06:31.921488 6918 event.go:377] Event(v1.ObjectRefere\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T10:06:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8v9j6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-75qcv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.505396 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"99b8492d-66ff-496d-8514-5d50a498e04a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e9dce7bdf9b34a32da9e47ee5c7823864c05c737beaeb598d9d49c8de44e31b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9c0b4b6d0f98425f08f3a6b47fb54aac3670f0bbd6f59a87dd5c40d8400e14a2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ad697ade69aea3061d955cd8b2ad8f9c0d33e7eda9357fb292d662f3434893d0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://069c6d5eda978abd7e64ac4f2e54612c78bf09efcd7bc9a2ef782e5a3227683f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9d9face62b93a7dafa45b1c3cbaf30967543bcfb6c8cfd9da85840de426d990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:40Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a3dbe761befaf063f590ea8b872799d1c40373e143701a2c8a1d9e74c5f994c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6b87bb12d0de856730f26468ddd876d6283d91e790518c17693cde2b1a71cb73\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T10:05:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T10:05:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-987lm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-j9tk4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.517774 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1130e920-d4a8-44fe-8bb4-050213d1cff2\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b96b6658fb4fce7a4e10dc94264c14d13eb805718968f2e45b20c36d4b2a6fce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T10:05:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nzq58\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:36Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-frh4j\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.529225 4923 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0ffdd04f-94e8-4fca-9676-ef87952debcc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T10:05:50Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgphv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T10:05:50Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-dxnsb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T10:06:38Z is after 2025-08-24T17:21:41Z" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.572910 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.572942 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.572951 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.572964 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.572972 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.601202 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:38 crc kubenswrapper[4923]: E1009 10:06:38.601788 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.676365 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.676932 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.677063 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.677184 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.677286 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.780837 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.780883 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.780892 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.780908 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.780918 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.884324 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.884379 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.884392 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.884412 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.884425 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.987792 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.987896 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.987919 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.987953 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:38 crc kubenswrapper[4923]: I1009 10:06:38.987980 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:38Z","lastTransitionTime":"2025-10-09T10:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.090610 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.090642 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.090651 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.090667 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.090675 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:39Z","lastTransitionTime":"2025-10-09T10:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.193016 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.193054 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.193064 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.193082 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.193097 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:39Z","lastTransitionTime":"2025-10-09T10:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.295441 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.295541 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.295557 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.295579 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.295594 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:39Z","lastTransitionTime":"2025-10-09T10:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.399500 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.399555 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.399565 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.399586 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.399598 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:39Z","lastTransitionTime":"2025-10-09T10:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.421370 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.421543 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.421612 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.421795 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.421708694 +0000 UTC m=+149.489890440 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.421822 4923 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.421883 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.421919 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.421941 4923 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.421962 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.422011 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.421985672 +0000 UTC m=+149.490167418 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.422048 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.422035493 +0000 UTC m=+149.490217519 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.422133 4923 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.422301 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.422227838 +0000 UTC m=+149.490409594 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.502890 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.503463 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.503547 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.503638 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.503736 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:39Z","lastTransitionTime":"2025-10-09T10:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.523530 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.523720 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.523774 4923 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.523789 4923 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.523858 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.523838977 +0000 UTC m=+149.592020733 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.600817 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.600852 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.600847 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.600970 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.601135 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:39 crc kubenswrapper[4923]: E1009 10:06:39.601216 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.606479 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.606527 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.606541 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.606556 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.606571 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:39Z","lastTransitionTime":"2025-10-09T10:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.709039 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.709121 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.709140 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.709172 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.709190 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:39Z","lastTransitionTime":"2025-10-09T10:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.811769 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.811808 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.811822 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.811838 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.811848 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:39Z","lastTransitionTime":"2025-10-09T10:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.915336 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.915390 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.915402 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.915423 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:39 crc kubenswrapper[4923]: I1009 10:06:39.915434 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:39Z","lastTransitionTime":"2025-10-09T10:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.018739 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.018794 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.018803 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.018819 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.018833 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.122303 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.122336 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.122344 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.122359 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.122371 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.224792 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.225066 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.225159 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.225266 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.225366 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.327039 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.327081 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.327092 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.327108 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.327119 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.429780 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.429842 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.429859 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.429885 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.429900 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.533160 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.533239 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.533252 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.533283 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.533297 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.601029 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:40 crc kubenswrapper[4923]: E1009 10:06:40.601157 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.635382 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.635623 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.635765 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.635872 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.635961 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.738560 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.738597 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.738613 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.738627 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.738638 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.841302 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.841332 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.841340 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.841352 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.841361 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.944331 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.944385 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.944396 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.944415 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:40 crc kubenswrapper[4923]: I1009 10:06:40.944428 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:40Z","lastTransitionTime":"2025-10-09T10:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.046246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.046317 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.046327 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.046346 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.046358 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.149180 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.149218 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.149231 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.149247 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.149260 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.251512 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.251837 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.251845 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.251863 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.251874 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.354885 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.354953 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.354966 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.354985 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.354998 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.461588 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.461634 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.461645 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.461662 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.461677 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.563886 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.563937 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.563948 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.563965 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.563976 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.600339 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.600440 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.600440 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:41 crc kubenswrapper[4923]: E1009 10:06:41.600547 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:41 crc kubenswrapper[4923]: E1009 10:06:41.600636 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:41 crc kubenswrapper[4923]: E1009 10:06:41.600730 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.666711 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.666770 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.666778 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.666793 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.666804 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.769748 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.769794 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.769802 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.769815 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.769827 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.872179 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.872226 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.872239 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.872262 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.872275 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.974913 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.975246 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.975334 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.975428 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:41 crc kubenswrapper[4923]: I1009 10:06:41.975519 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:41Z","lastTransitionTime":"2025-10-09T10:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.078167 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.078218 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.078229 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.078247 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.078258 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:42Z","lastTransitionTime":"2025-10-09T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.181048 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.181092 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.181102 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.181121 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.181131 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:42Z","lastTransitionTime":"2025-10-09T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.283326 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.283369 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.283378 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.283394 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.283403 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:42Z","lastTransitionTime":"2025-10-09T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.386558 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.386604 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.386616 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.386633 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.386644 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:42Z","lastTransitionTime":"2025-10-09T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.489109 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.489146 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.489175 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.489192 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.489202 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:42Z","lastTransitionTime":"2025-10-09T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.592435 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.592480 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.592491 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.592508 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.592518 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:42Z","lastTransitionTime":"2025-10-09T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.601061 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:42 crc kubenswrapper[4923]: E1009 10:06:42.601247 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.694640 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.694682 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.694693 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.694708 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.694722 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:42Z","lastTransitionTime":"2025-10-09T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.797317 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.797362 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.797373 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.797389 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.797404 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:42Z","lastTransitionTime":"2025-10-09T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.899712 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.899745 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.899771 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.899785 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:42 crc kubenswrapper[4923]: I1009 10:06:42.899794 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:42Z","lastTransitionTime":"2025-10-09T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.002549 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.002598 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.002611 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.002645 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.002658 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:43Z","lastTransitionTime":"2025-10-09T10:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.035216 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.035260 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.035271 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.035289 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.035301 4923 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T10:06:43Z","lastTransitionTime":"2025-10-09T10:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.076274 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6"] Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.076898 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.078738 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.078976 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.079207 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.079377 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.120877 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-4cn4f" podStartSLOduration=67.120855036 podStartE2EDuration="1m7.120855036s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:43.120451875 +0000 UTC m=+89.188633651" watchObservedRunningTime="2025-10-09 10:06:43.120855036 +0000 UTC m=+89.189036792" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.150042 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bxn56" podStartSLOduration=67.150026843 podStartE2EDuration="1m7.150026843s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:43.134418701 +0000 UTC m=+89.202600447" watchObservedRunningTime="2025-10-09 10:06:43.150026843 +0000 UTC m=+89.218208599" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.158037 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.158091 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.158109 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.158186 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.158208 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.164737 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=69.164720559 podStartE2EDuration="1m9.164720559s" podCreationTimestamp="2025-10-09 10:05:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:43.151329839 +0000 UTC m=+89.219511595" watchObservedRunningTime="2025-10-09 10:06:43.164720559 +0000 UTC m=+89.232902315" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.173718 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-k72w6" podStartSLOduration=67.173700217 podStartE2EDuration="1m7.173700217s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:43.173519472 +0000 UTC m=+89.241701228" watchObservedRunningTime="2025-10-09 10:06:43.173700217 +0000 UTC m=+89.241881973" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.205062 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=68.205044724 podStartE2EDuration="1m8.205044724s" podCreationTimestamp="2025-10-09 10:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:43.191340535 +0000 UTC m=+89.259522311" watchObservedRunningTime="2025-10-09 10:06:43.205044724 +0000 UTC m=+89.273226480" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.227907 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-nrck8" podStartSLOduration=67.227885496 podStartE2EDuration="1m7.227885496s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:43.216409319 +0000 UTC m=+89.284591075" watchObservedRunningTime="2025-10-09 10:06:43.227885496 +0000 UTC m=+89.296067252" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.245320 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=34.245287377 podStartE2EDuration="34.245287377s" podCreationTimestamp="2025-10-09 10:06:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:43.245282047 +0000 UTC m=+89.313463813" watchObservedRunningTime="2025-10-09 10:06:43.245287377 +0000 UTC m=+89.313469143" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.258795 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.258860 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.258884 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.258904 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.258963 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.258992 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.259043 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.259887 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-service-ca\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.273450 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.286501 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae6a3411-1a01-4181-aaa9-1e3d31bf7969-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-ctvq6\" (UID: \"ae6a3411-1a01-4181-aaa9-1e3d31bf7969\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.320352 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-j9tk4" podStartSLOduration=67.320298621 podStartE2EDuration="1m7.320298621s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:43.320198518 +0000 UTC m=+89.388380284" watchObservedRunningTime="2025-10-09 10:06:43.320298621 +0000 UTC m=+89.388480377" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.334561 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podStartSLOduration=67.334539064 podStartE2EDuration="1m7.334539064s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:43.334258247 +0000 UTC m=+89.402440003" watchObservedRunningTime="2025-10-09 10:06:43.334539064 +0000 UTC m=+89.402720830" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.390466 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" Oct 09 10:06:43 crc kubenswrapper[4923]: W1009 10:06:43.404044 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae6a3411_1a01_4181_aaa9_1e3d31bf7969.slice/crio-a3f0e3de3cd8c33604f264f9e162969941373abe457702f9a60c6dc0853bd2e4 WatchSource:0}: Error finding container a3f0e3de3cd8c33604f264f9e162969941373abe457702f9a60c6dc0853bd2e4: Status 404 returned error can't find the container with id a3f0e3de3cd8c33604f264f9e162969941373abe457702f9a60c6dc0853bd2e4 Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.600617 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.600707 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:43 crc kubenswrapper[4923]: I1009 10:06:43.600618 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:43 crc kubenswrapper[4923]: E1009 10:06:43.600781 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:43 crc kubenswrapper[4923]: E1009 10:06:43.600858 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:43 crc kubenswrapper[4923]: E1009 10:06:43.600994 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:44 crc kubenswrapper[4923]: I1009 10:06:44.054660 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" event={"ID":"ae6a3411-1a01-4181-aaa9-1e3d31bf7969","Type":"ContainerStarted","Data":"a6e972450749eb645762839765d8b8ed81288766bd0e1a75b755cd9adb2e17fc"} Oct 09 10:06:44 crc kubenswrapper[4923]: I1009 10:06:44.054709 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" event={"ID":"ae6a3411-1a01-4181-aaa9-1e3d31bf7969","Type":"ContainerStarted","Data":"a3f0e3de3cd8c33604f264f9e162969941373abe457702f9a60c6dc0853bd2e4"} Oct 09 10:06:44 crc kubenswrapper[4923]: I1009 10:06:44.072162 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-ctvq6" podStartSLOduration=68.072148147 podStartE2EDuration="1m8.072148147s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:44.071538201 +0000 UTC m=+90.139719957" watchObservedRunningTime="2025-10-09 10:06:44.072148147 +0000 UTC m=+90.140329893" Oct 09 10:06:44 crc kubenswrapper[4923]: I1009 10:06:44.600360 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:44 crc kubenswrapper[4923]: E1009 10:06:44.601349 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:45 crc kubenswrapper[4923]: I1009 10:06:45.601107 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:45 crc kubenswrapper[4923]: I1009 10:06:45.601251 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:45 crc kubenswrapper[4923]: E1009 10:06:45.601360 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:45 crc kubenswrapper[4923]: I1009 10:06:45.601394 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:45 crc kubenswrapper[4923]: E1009 10:06:45.601566 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:45 crc kubenswrapper[4923]: E1009 10:06:45.601630 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:46 crc kubenswrapper[4923]: I1009 10:06:46.600484 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:46 crc kubenswrapper[4923]: E1009 10:06:46.600607 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:47 crc kubenswrapper[4923]: I1009 10:06:47.600655 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:47 crc kubenswrapper[4923]: I1009 10:06:47.600725 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:47 crc kubenswrapper[4923]: E1009 10:06:47.600828 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:47 crc kubenswrapper[4923]: I1009 10:06:47.600979 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:47 crc kubenswrapper[4923]: E1009 10:06:47.601127 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:47 crc kubenswrapper[4923]: E1009 10:06:47.601341 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:48 crc kubenswrapper[4923]: I1009 10:06:48.601364 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:48 crc kubenswrapper[4923]: E1009 10:06:48.602186 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:48 crc kubenswrapper[4923]: I1009 10:06:48.620645 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 09 10:06:49 crc kubenswrapper[4923]: I1009 10:06:49.600663 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:49 crc kubenswrapper[4923]: I1009 10:06:49.600823 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:49 crc kubenswrapper[4923]: E1009 10:06:49.601423 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:49 crc kubenswrapper[4923]: I1009 10:06:49.600854 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:49 crc kubenswrapper[4923]: E1009 10:06:49.601502 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:49 crc kubenswrapper[4923]: E1009 10:06:49.601780 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:50 crc kubenswrapper[4923]: I1009 10:06:50.603446 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:50 crc kubenswrapper[4923]: E1009 10:06:50.603684 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:50 crc kubenswrapper[4923]: I1009 10:06:50.605961 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:06:50 crc kubenswrapper[4923]: E1009 10:06:50.606214 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\"" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" Oct 09 10:06:51 crc kubenswrapper[4923]: I1009 10:06:51.600611 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:51 crc kubenswrapper[4923]: I1009 10:06:51.600614 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:51 crc kubenswrapper[4923]: E1009 10:06:51.600839 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:51 crc kubenswrapper[4923]: I1009 10:06:51.600634 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:51 crc kubenswrapper[4923]: E1009 10:06:51.600922 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:51 crc kubenswrapper[4923]: E1009 10:06:51.601008 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:52 crc kubenswrapper[4923]: I1009 10:06:52.600992 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:52 crc kubenswrapper[4923]: E1009 10:06:52.601196 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:53 crc kubenswrapper[4923]: I1009 10:06:53.601276 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:53 crc kubenswrapper[4923]: I1009 10:06:53.602005 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:53 crc kubenswrapper[4923]: E1009 10:06:53.602043 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:53 crc kubenswrapper[4923]: I1009 10:06:53.601306 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:53 crc kubenswrapper[4923]: E1009 10:06:53.602219 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:53 crc kubenswrapper[4923]: E1009 10:06:53.602276 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:54 crc kubenswrapper[4923]: I1009 10:06:54.600674 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:54 crc kubenswrapper[4923]: E1009 10:06:54.602185 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:54 crc kubenswrapper[4923]: I1009 10:06:54.617236 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=6.617212423 podStartE2EDuration="6.617212423s" podCreationTimestamp="2025-10-09 10:06:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:06:54.616538434 +0000 UTC m=+100.684720190" watchObservedRunningTime="2025-10-09 10:06:54.617212423 +0000 UTC m=+100.685394179" Oct 09 10:06:54 crc kubenswrapper[4923]: I1009 10:06:54.685527 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:54 crc kubenswrapper[4923]: E1009 10:06:54.685717 4923 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:06:54 crc kubenswrapper[4923]: E1009 10:06:54.685831 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs podName:0ffdd04f-94e8-4fca-9676-ef87952debcc nodeName:}" failed. No retries permitted until 2025-10-09 10:07:58.68580547 +0000 UTC m=+164.753987236 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs") pod "network-metrics-daemon-dxnsb" (UID: "0ffdd04f-94e8-4fca-9676-ef87952debcc") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 10:06:55 crc kubenswrapper[4923]: I1009 10:06:55.600868 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:55 crc kubenswrapper[4923]: I1009 10:06:55.600910 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:55 crc kubenswrapper[4923]: E1009 10:06:55.600996 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:55 crc kubenswrapper[4923]: I1009 10:06:55.601054 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:55 crc kubenswrapper[4923]: E1009 10:06:55.601196 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:55 crc kubenswrapper[4923]: E1009 10:06:55.601315 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:56 crc kubenswrapper[4923]: I1009 10:06:56.600510 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:56 crc kubenswrapper[4923]: E1009 10:06:56.600677 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:57 crc kubenswrapper[4923]: I1009 10:06:57.600987 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:57 crc kubenswrapper[4923]: I1009 10:06:57.601049 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:57 crc kubenswrapper[4923]: I1009 10:06:57.601057 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:57 crc kubenswrapper[4923]: E1009 10:06:57.601250 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:57 crc kubenswrapper[4923]: E1009 10:06:57.601444 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:06:57 crc kubenswrapper[4923]: E1009 10:06:57.601873 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:58 crc kubenswrapper[4923]: I1009 10:06:58.601041 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:06:58 crc kubenswrapper[4923]: E1009 10:06:58.601177 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:06:59 crc kubenswrapper[4923]: I1009 10:06:59.600540 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:06:59 crc kubenswrapper[4923]: I1009 10:06:59.600578 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:06:59 crc kubenswrapper[4923]: I1009 10:06:59.600553 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:06:59 crc kubenswrapper[4923]: E1009 10:06:59.600695 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:06:59 crc kubenswrapper[4923]: E1009 10:06:59.600813 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:06:59 crc kubenswrapper[4923]: E1009 10:06:59.600876 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:00 crc kubenswrapper[4923]: I1009 10:07:00.600729 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:00 crc kubenswrapper[4923]: E1009 10:07:00.601049 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:01 crc kubenswrapper[4923]: I1009 10:07:01.601168 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:01 crc kubenswrapper[4923]: I1009 10:07:01.601202 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:01 crc kubenswrapper[4923]: I1009 10:07:01.601182 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:01 crc kubenswrapper[4923]: E1009 10:07:01.601504 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:01 crc kubenswrapper[4923]: E1009 10:07:01.601631 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:01 crc kubenswrapper[4923]: E1009 10:07:01.601694 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:02 crc kubenswrapper[4923]: I1009 10:07:02.600592 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:02 crc kubenswrapper[4923]: E1009 10:07:02.600843 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:02 crc kubenswrapper[4923]: I1009 10:07:02.615896 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 09 10:07:03 crc kubenswrapper[4923]: I1009 10:07:03.601059 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:03 crc kubenswrapper[4923]: I1009 10:07:03.601114 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:03 crc kubenswrapper[4923]: E1009 10:07:03.601475 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:03 crc kubenswrapper[4923]: I1009 10:07:03.601132 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:03 crc kubenswrapper[4923]: E1009 10:07:03.601805 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:03 crc kubenswrapper[4923]: E1009 10:07:03.601975 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:03 crc kubenswrapper[4923]: I1009 10:07:03.602061 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:07:03 crc kubenswrapper[4923]: E1009 10:07:03.602291 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-75qcv_openshift-ovn-kubernetes(bc2a175c-508c-4f50-8205-013a9482c989)\"" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" Oct 09 10:07:04 crc kubenswrapper[4923]: I1009 10:07:04.602806 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:04 crc kubenswrapper[4923]: E1009 10:07:04.603585 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:05 crc kubenswrapper[4923]: I1009 10:07:05.601214 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:05 crc kubenswrapper[4923]: I1009 10:07:05.601234 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:05 crc kubenswrapper[4923]: I1009 10:07:05.601234 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:05 crc kubenswrapper[4923]: E1009 10:07:05.601494 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:05 crc kubenswrapper[4923]: E1009 10:07:05.601558 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:05 crc kubenswrapper[4923]: E1009 10:07:05.601348 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:06 crc kubenswrapper[4923]: I1009 10:07:06.601603 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:06 crc kubenswrapper[4923]: E1009 10:07:06.601792 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:07 crc kubenswrapper[4923]: I1009 10:07:07.600994 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:07 crc kubenswrapper[4923]: I1009 10:07:07.600988 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:07 crc kubenswrapper[4923]: I1009 10:07:07.601128 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:07 crc kubenswrapper[4923]: E1009 10:07:07.601288 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:07 crc kubenswrapper[4923]: E1009 10:07:07.601517 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:07 crc kubenswrapper[4923]: E1009 10:07:07.601568 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:08 crc kubenswrapper[4923]: I1009 10:07:08.600910 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:08 crc kubenswrapper[4923]: E1009 10:07:08.601343 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:09 crc kubenswrapper[4923]: I1009 10:07:09.600708 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:09 crc kubenswrapper[4923]: I1009 10:07:09.600833 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:09 crc kubenswrapper[4923]: I1009 10:07:09.600794 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:09 crc kubenswrapper[4923]: E1009 10:07:09.600972 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:09 crc kubenswrapper[4923]: E1009 10:07:09.601104 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:09 crc kubenswrapper[4923]: E1009 10:07:09.601217 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:10 crc kubenswrapper[4923]: I1009 10:07:10.601394 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:10 crc kubenswrapper[4923]: E1009 10:07:10.601686 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.134568 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/1.log" Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.135157 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/0.log" Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.135188 4923 generic.go:334] "Generic (PLEG): container finished" podID="d5c5f455-4e94-4f9a-b6fa-2535685c6354" containerID="29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a" exitCode=1 Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.135258 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4cn4f" event={"ID":"d5c5f455-4e94-4f9a-b6fa-2535685c6354","Type":"ContainerDied","Data":"29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a"} Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.135296 4923 scope.go:117] "RemoveContainer" containerID="c14704f9e50f1540cd6ee82b936510cff8a2a958ba303799bd401b1b09e63a24" Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.135628 4923 scope.go:117] "RemoveContainer" containerID="29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a" Oct 09 10:07:11 crc kubenswrapper[4923]: E1009 10:07:11.135904 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-4cn4f_openshift-multus(d5c5f455-4e94-4f9a-b6fa-2535685c6354)\"" pod="openshift-multus/multus-4cn4f" podUID="d5c5f455-4e94-4f9a-b6fa-2535685c6354" Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.165662 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=9.165626434 podStartE2EDuration="9.165626434s" podCreationTimestamp="2025-10-09 10:07:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:04.643153025 +0000 UTC m=+110.711334841" watchObservedRunningTime="2025-10-09 10:07:11.165626434 +0000 UTC m=+117.233808190" Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.600405 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.600449 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:11 crc kubenswrapper[4923]: I1009 10:07:11.600423 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:11 crc kubenswrapper[4923]: E1009 10:07:11.600598 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:11 crc kubenswrapper[4923]: E1009 10:07:11.600718 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:11 crc kubenswrapper[4923]: E1009 10:07:11.600848 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:12 crc kubenswrapper[4923]: I1009 10:07:12.142258 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/1.log" Oct 09 10:07:12 crc kubenswrapper[4923]: I1009 10:07:12.600546 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:12 crc kubenswrapper[4923]: E1009 10:07:12.600831 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:13 crc kubenswrapper[4923]: I1009 10:07:13.600856 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:13 crc kubenswrapper[4923]: I1009 10:07:13.600921 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:13 crc kubenswrapper[4923]: E1009 10:07:13.601060 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:13 crc kubenswrapper[4923]: I1009 10:07:13.600921 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:13 crc kubenswrapper[4923]: E1009 10:07:13.601258 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:13 crc kubenswrapper[4923]: E1009 10:07:13.601374 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:14 crc kubenswrapper[4923]: I1009 10:07:14.600564 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:14 crc kubenswrapper[4923]: E1009 10:07:14.601579 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:14 crc kubenswrapper[4923]: E1009 10:07:14.611715 4923 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 09 10:07:14 crc kubenswrapper[4923]: E1009 10:07:14.701542 4923 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 09 10:07:15 crc kubenswrapper[4923]: I1009 10:07:15.600297 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:15 crc kubenswrapper[4923]: I1009 10:07:15.600427 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:15 crc kubenswrapper[4923]: I1009 10:07:15.600427 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:15 crc kubenswrapper[4923]: E1009 10:07:15.601290 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:15 crc kubenswrapper[4923]: E1009 10:07:15.601044 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:15 crc kubenswrapper[4923]: E1009 10:07:15.601441 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:16 crc kubenswrapper[4923]: I1009 10:07:16.601462 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:16 crc kubenswrapper[4923]: E1009 10:07:16.601700 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:17 crc kubenswrapper[4923]: I1009 10:07:17.600570 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:17 crc kubenswrapper[4923]: I1009 10:07:17.600636 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:17 crc kubenswrapper[4923]: E1009 10:07:17.600873 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:17 crc kubenswrapper[4923]: I1009 10:07:17.600923 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:17 crc kubenswrapper[4923]: E1009 10:07:17.601444 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:17 crc kubenswrapper[4923]: E1009 10:07:17.601659 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:17 crc kubenswrapper[4923]: I1009 10:07:17.602002 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:07:18 crc kubenswrapper[4923]: I1009 10:07:18.166677 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/3.log" Oct 09 10:07:18 crc kubenswrapper[4923]: I1009 10:07:18.168705 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerStarted","Data":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} Oct 09 10:07:18 crc kubenswrapper[4923]: I1009 10:07:18.170027 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:07:18 crc kubenswrapper[4923]: I1009 10:07:18.208262 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podStartSLOduration=102.208238354 podStartE2EDuration="1m42.208238354s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:18.208133371 +0000 UTC m=+124.276315127" watchObservedRunningTime="2025-10-09 10:07:18.208238354 +0000 UTC m=+124.276420110" Oct 09 10:07:18 crc kubenswrapper[4923]: I1009 10:07:18.537347 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-dxnsb"] Oct 09 10:07:18 crc kubenswrapper[4923]: I1009 10:07:18.537525 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:18 crc kubenswrapper[4923]: E1009 10:07:18.537642 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:19 crc kubenswrapper[4923]: I1009 10:07:19.600594 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:19 crc kubenswrapper[4923]: I1009 10:07:19.600668 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:19 crc kubenswrapper[4923]: E1009 10:07:19.601294 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:19 crc kubenswrapper[4923]: I1009 10:07:19.600737 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:19 crc kubenswrapper[4923]: I1009 10:07:19.600702 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:19 crc kubenswrapper[4923]: E1009 10:07:19.601472 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:19 crc kubenswrapper[4923]: E1009 10:07:19.601561 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:19 crc kubenswrapper[4923]: E1009 10:07:19.601838 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:19 crc kubenswrapper[4923]: E1009 10:07:19.703409 4923 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 09 10:07:21 crc kubenswrapper[4923]: I1009 10:07:21.601290 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:21 crc kubenswrapper[4923]: E1009 10:07:21.601447 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:21 crc kubenswrapper[4923]: I1009 10:07:21.601645 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:21 crc kubenswrapper[4923]: E1009 10:07:21.601732 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:21 crc kubenswrapper[4923]: I1009 10:07:21.601742 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:21 crc kubenswrapper[4923]: I1009 10:07:21.601832 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:21 crc kubenswrapper[4923]: E1009 10:07:21.601910 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:21 crc kubenswrapper[4923]: E1009 10:07:21.602199 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:23 crc kubenswrapper[4923]: I1009 10:07:23.601196 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:23 crc kubenswrapper[4923]: I1009 10:07:23.601415 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:23 crc kubenswrapper[4923]: E1009 10:07:23.601458 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:23 crc kubenswrapper[4923]: I1009 10:07:23.601497 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:23 crc kubenswrapper[4923]: I1009 10:07:23.601416 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:23 crc kubenswrapper[4923]: E1009 10:07:23.601667 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:23 crc kubenswrapper[4923]: E1009 10:07:23.601801 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:23 crc kubenswrapper[4923]: I1009 10:07:23.602029 4923 scope.go:117] "RemoveContainer" containerID="29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a" Oct 09 10:07:23 crc kubenswrapper[4923]: E1009 10:07:23.602101 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:24 crc kubenswrapper[4923]: I1009 10:07:24.198230 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/1.log" Oct 09 10:07:24 crc kubenswrapper[4923]: I1009 10:07:24.198305 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4cn4f" event={"ID":"d5c5f455-4e94-4f9a-b6fa-2535685c6354","Type":"ContainerStarted","Data":"8aa38bf09b127046fbe3053995133ff9c81ee52e4a4cd7e17444e49b1e6ce556"} Oct 09 10:07:24 crc kubenswrapper[4923]: E1009 10:07:24.704172 4923 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 09 10:07:25 crc kubenswrapper[4923]: I1009 10:07:25.600925 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:25 crc kubenswrapper[4923]: I1009 10:07:25.600934 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:25 crc kubenswrapper[4923]: E1009 10:07:25.601420 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:25 crc kubenswrapper[4923]: I1009 10:07:25.601020 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:25 crc kubenswrapper[4923]: E1009 10:07:25.601476 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:25 crc kubenswrapper[4923]: I1009 10:07:25.600961 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:25 crc kubenswrapper[4923]: E1009 10:07:25.601504 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:25 crc kubenswrapper[4923]: E1009 10:07:25.601568 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:27 crc kubenswrapper[4923]: I1009 10:07:27.600837 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:27 crc kubenswrapper[4923]: I1009 10:07:27.600931 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:27 crc kubenswrapper[4923]: I1009 10:07:27.600939 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:27 crc kubenswrapper[4923]: I1009 10:07:27.600844 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:27 crc kubenswrapper[4923]: E1009 10:07:27.601104 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:27 crc kubenswrapper[4923]: E1009 10:07:27.601232 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:27 crc kubenswrapper[4923]: E1009 10:07:27.601330 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:27 crc kubenswrapper[4923]: E1009 10:07:27.601454 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:29 crc kubenswrapper[4923]: I1009 10:07:29.601039 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:29 crc kubenswrapper[4923]: I1009 10:07:29.601083 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:29 crc kubenswrapper[4923]: E1009 10:07:29.601202 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 10:07:29 crc kubenswrapper[4923]: I1009 10:07:29.601039 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:29 crc kubenswrapper[4923]: E1009 10:07:29.601328 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 10:07:29 crc kubenswrapper[4923]: E1009 10:07:29.601432 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 10:07:29 crc kubenswrapper[4923]: I1009 10:07:29.601517 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:29 crc kubenswrapper[4923]: E1009 10:07:29.601674 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-dxnsb" podUID="0ffdd04f-94e8-4fca-9676-ef87952debcc" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.601137 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.601230 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.601171 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.601138 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.604128 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.604137 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.604194 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.604599 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.604668 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 09 10:07:31 crc kubenswrapper[4923]: I1009 10:07:31.604841 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.538594 4923 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.584831 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tnbft"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.586278 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.591557 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.592192 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.593915 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-66f2z"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.594652 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.595285 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.595594 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.596390 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.596916 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-9rq8l"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.597308 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.597602 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.597663 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.598192 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9rq8l" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.601536 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.601776 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.601938 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.602199 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.602349 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.604747 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.609827 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.610638 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.610927 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.611119 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.611349 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.611445 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.611575 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.611844 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.617120 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9sd7w"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.617620 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.618014 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.618267 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-26ghk"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.618517 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.618896 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5qd2g"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.619141 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.619553 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dtw6g"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.620280 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.620580 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lwzgh"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.620908 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.621491 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.622075 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.622396 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.622696 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.623092 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.623373 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.623907 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.623991 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.624331 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.624714 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.624977 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.625248 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.628462 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.628648 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.629819 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.635851 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2dxwn"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.636679 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-44mh6"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.637098 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.637164 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.637502 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.637715 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.659957 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.660240 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.660460 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.661455 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.661853 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.662434 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.662468 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.665691 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-85h6b"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.666370 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.679115 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.679991 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.680613 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx26v\" (UniqueName: \"kubernetes.io/projected/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-kube-api-access-nx26v\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.680740 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-config\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.680860 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c37f95a3-8ee6-4cd3-bfd5-79f688335b19-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qw297\" (UID: \"c37f95a3-8ee6-4cd3-bfd5-79f688335b19\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.680939 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-oauth-serving-cert\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681044 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-service-ca\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681146 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh754\" (UniqueName: \"kubernetes.io/projected/a8c487fa-3436-4cc9-83b3-de49256e97c9-kube-api-access-bh754\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681215 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681315 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dz5nz\" (UniqueName: \"kubernetes.io/projected/c37f95a3-8ee6-4cd3-bfd5-79f688335b19-kube-api-access-dz5nz\") pod \"openshift-controller-manager-operator-756b6f6bc6-qw297\" (UID: \"c37f95a3-8ee6-4cd3-bfd5-79f688335b19\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681405 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-oauth-config\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681535 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-trusted-ca-bundle\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681625 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c37f95a3-8ee6-4cd3-bfd5-79f688335b19-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qw297\" (UID: \"c37f95a3-8ee6-4cd3-bfd5-79f688335b19\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681713 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-serving-cert\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681805 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-client-ca\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681907 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-config\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681993 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-serving-cert\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.682813 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.683894 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.681897 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.687304 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.682047 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.682169 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.682189 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.683117 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.684097 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.684237 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.684309 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.684335 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.684415 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.684921 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.686552 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.686604 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.686613 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.686683 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.688382 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.686721 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.686875 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.686938 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.689154 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.689240 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.689683 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.689841 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.690104 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.690300 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.690631 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.691144 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.691375 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5qc5h"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.692463 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.692569 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.692711 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.693209 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.692797 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.692843 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.692884 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.692918 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.692945 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.693006 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.693050 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.693164 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.693208 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.693246 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.695760 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.695938 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.695945 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.696341 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697247 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697407 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697512 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697610 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697634 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697713 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697766 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697781 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697818 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697864 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697913 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.697959 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698002 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698043 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698103 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698129 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698220 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698301 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698395 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698403 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698495 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698002 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698820 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.698953 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.699461 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.700564 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.702062 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.702171 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.704629 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.710248 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tnbft"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.712005 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.712777 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.735501 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.736267 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.736705 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.736968 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvsh5"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.739014 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.740667 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.758697 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.759672 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.761251 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.762422 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.762461 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-897kb"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.763164 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.766449 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.766656 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.767161 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.768183 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.770943 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.771881 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.773875 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.774492 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b9mj6"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.775868 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.776266 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.774799 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.776652 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.776786 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.777156 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.779831 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-66f2z"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.780214 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.781792 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-8n9cm"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782502 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cecd9316-6fa9-4f5f-bc1c-aa468860ea3d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wb9v2\" (UID: \"cecd9316-6fa9-4f5f-bc1c-aa468860ea3d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782544 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782569 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8b9202a-16ae-4569-9b82-e7342d6b82c5-config\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782626 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28996\" (UniqueName: \"kubernetes.io/projected/981e0fe0-3fa9-403a-a147-5d31881cf83f-kube-api-access-28996\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782644 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xjcb\" (UniqueName: \"kubernetes.io/projected/54841e96-db7c-4118-bbfb-3d35bbf59a4d-kube-api-access-2xjcb\") pod \"dns-operator-744455d44c-2dxwn\" (UID: \"54841e96-db7c-4118-bbfb-3d35bbf59a4d\") " pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782663 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcpkj\" (UniqueName: \"kubernetes.io/projected/3bf69a9c-5575-4c6f-8751-5ca9f3640863-kube-api-access-wcpkj\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvfh6\" (UID: \"3bf69a9c-5575-4c6f-8751-5ca9f3640863\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782682 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tp52\" (UniqueName: \"kubernetes.io/projected/a8e9f6b9-0e46-4271-89c3-657726d9dbed-kube-api-access-4tp52\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782699 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-serving-cert\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782732 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/981e0fe0-3fa9-403a-a147-5d31881cf83f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782770 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47zpm\" (UniqueName: \"kubernetes.io/projected/8abb0401-2b84-4ae7-abe1-6e151c6b1bb6-kube-api-access-47zpm\") pod \"catalog-operator-68c6474976-v2b95\" (UID: \"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782796 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4t64\" (UniqueName: \"kubernetes.io/projected/d755342a-48bb-467b-bc8a-5f18a8389ceb-kube-api-access-r4t64\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782813 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-policies\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782831 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpvmn\" (UniqueName: \"kubernetes.io/projected/401d0883-625c-4bb6-80ba-8471f93c388b-kube-api-access-mpvmn\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782860 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8e9f6b9-0e46-4271-89c3-657726d9dbed-serving-cert\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782884 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ssrd\" (UniqueName: \"kubernetes.io/projected/45725226-1710-4204-8055-ecea69f828ee-kube-api-access-4ssrd\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782894 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782907 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-config\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.782962 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8abb0401-2b84-4ae7-abe1-6e151c6b1bb6-srv-cert\") pod \"catalog-operator-68c6474976-v2b95\" (UID: \"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783004 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-config\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783043 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-serving-cert\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783073 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/68e9fb5c-167b-4528-9c31-97cbdc41aecd-stats-auth\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783092 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkbtn\" (UniqueName: \"kubernetes.io/projected/68e9fb5c-167b-4528-9c31-97cbdc41aecd-kube-api-access-tkbtn\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783117 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a8e9f6b9-0e46-4271-89c3-657726d9dbed-trusted-ca\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783172 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b0d7439-9f5b-4a4f-ab08-64d77a966845-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zrf6j\" (UID: \"9b0d7439-9f5b-4a4f-ab08-64d77a966845\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783208 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/401d0883-625c-4bb6-80ba-8471f93c388b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783229 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d755342a-48bb-467b-bc8a-5f18a8389ceb-auth-proxy-config\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783249 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783271 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx26v\" (UniqueName: \"kubernetes.io/projected/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-kube-api-access-nx26v\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783289 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d-serving-cert\") pod \"openshift-config-operator-7777fb866f-vmrkj\" (UID: \"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783309 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4464ee82-fef7-46d6-8118-1e1d685e63f6-node-pullsecrets\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783328 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783348 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5qc5h\" (UID: \"6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783367 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7f7v\" (UniqueName: \"kubernetes.io/projected/cecd9316-6fa9-4f5f-bc1c-aa468860ea3d-kube-api-access-j7f7v\") pod \"cluster-samples-operator-665b6dd947-wb9v2\" (UID: \"cecd9316-6fa9-4f5f-bc1c-aa468860ea3d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783383 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/401d0883-625c-4bb6-80ba-8471f93c388b-audit-policies\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783404 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/68e9fb5c-167b-4528-9c31-97cbdc41aecd-service-ca-bundle\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783419 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qsgx\" (UniqueName: \"kubernetes.io/projected/7eebe686-79b8-4742-bd0b-09db708418d8-kube-api-access-5qsgx\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783437 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89w7l\" (UniqueName: \"kubernetes.io/projected/a8b9202a-16ae-4569-9b82-e7342d6b82c5-kube-api-access-89w7l\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783457 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4464ee82-fef7-46d6-8118-1e1d685e63f6-audit-dir\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783472 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8e9f6b9-0e46-4271-89c3-657726d9dbed-config\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783491 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-config\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783508 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7eebe686-79b8-4742-bd0b-09db708418d8-trusted-ca\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783526 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz9j5\" (UniqueName: \"kubernetes.io/projected/d06b0170-f718-4285-a25b-b5c07868e3c2-kube-api-access-jz9j5\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783546 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c37f95a3-8ee6-4cd3-bfd5-79f688335b19-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qw297\" (UID: \"c37f95a3-8ee6-4cd3-bfd5-79f688335b19\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783562 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-client-ca\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783581 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783597 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d06b0170-f718-4285-a25b-b5c07868e3c2-config\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783615 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-oauth-serving-cert\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783633 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mw6r9\" (UniqueName: \"kubernetes.io/projected/6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0-kube-api-access-mw6r9\") pod \"multus-admission-controller-857f4d67dd-5qc5h\" (UID: \"6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783648 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d755342a-48bb-467b-bc8a-5f18a8389ceb-config\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783665 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783686 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/401d0883-625c-4bb6-80ba-8471f93c388b-serving-cert\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783702 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/401d0883-625c-4bb6-80ba-8471f93c388b-audit-dir\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783720 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d06b0170-f718-4285-a25b-b5c07868e3c2-images\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783736 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a8b9202a-16ae-4569-9b82-e7342d6b82c5-etcd-client\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783773 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b0d7439-9f5b-4a4f-ab08-64d77a966845-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zrf6j\" (UID: \"9b0d7439-9f5b-4a4f-ab08-64d77a966845\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783791 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783807 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxwqh\" (UniqueName: \"kubernetes.io/projected/c37f83cf-f85a-4872-8e9e-ed37253c753f-kube-api-access-zxwqh\") pod \"control-plane-machine-set-operator-78cbb6b69f-z97pw\" (UID: \"c37f83cf-f85a-4872-8e9e-ed37253c753f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783825 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-service-ca\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783841 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783859 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7eebe686-79b8-4742-bd0b-09db708418d8-metrics-tls\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783876 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-audit\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783893 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7qf5\" (UniqueName: \"kubernetes.io/projected/9b0d7439-9f5b-4a4f-ab08-64d77a966845-kube-api-access-z7qf5\") pod \"openshift-apiserver-operator-796bbdcf4f-zrf6j\" (UID: \"9b0d7439-9f5b-4a4f-ab08-64d77a966845\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783910 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/981e0fe0-3fa9-403a-a147-5d31881cf83f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783926 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjcx4\" (UniqueName: \"kubernetes.io/projected/dc7bae79-3278-46f0-91fd-9a49b4057200-kube-api-access-rjcx4\") pod \"downloads-7954f5f757-9rq8l\" (UID: \"dc7bae79-3278-46f0-91fd-9a49b4057200\") " pod="openshift-console/downloads-7954f5f757-9rq8l" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783955 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh754\" (UniqueName: \"kubernetes.io/projected/a8c487fa-3436-4cc9-83b3-de49256e97c9-kube-api-access-bh754\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783974 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.783990 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n56qw\" (UniqueName: \"kubernetes.io/projected/c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d-kube-api-access-n56qw\") pod \"openshift-config-operator-7777fb866f-vmrkj\" (UID: \"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784392 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68e9fb5c-167b-4528-9c31-97cbdc41aecd-metrics-certs\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784512 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45725226-1710-4204-8055-ecea69f828ee-serving-cert\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784572 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dz5nz\" (UniqueName: \"kubernetes.io/projected/c37f95a3-8ee6-4cd3-bfd5-79f688335b19-kube-api-access-dz5nz\") pod \"openshift-controller-manager-operator-756b6f6bc6-qw297\" (UID: \"c37f95a3-8ee6-4cd3-bfd5-79f688335b19\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784597 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784643 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-config\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784680 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/54841e96-db7c-4118-bbfb-3d35bbf59a4d-metrics-tls\") pod \"dns-operator-744455d44c-2dxwn\" (UID: \"54841e96-db7c-4118-bbfb-3d35bbf59a4d\") " pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784719 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784738 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7eebe686-79b8-4742-bd0b-09db708418d8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784783 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784800 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784958 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8n9cm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.784802 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/68e9fb5c-167b-4528-9c31-97cbdc41aecd-default-certificate\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785065 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d06b0170-f718-4285-a25b-b5c07868e3c2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785120 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bf69a9c-5575-4c6f-8751-5ca9f3640863-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvfh6\" (UID: \"3bf69a9c-5575-4c6f-8751-5ca9f3640863\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785153 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phtw6\" (UniqueName: \"kubernetes.io/projected/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-kube-api-access-phtw6\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785171 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785193 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/401d0883-625c-4bb6-80ba-8471f93c388b-etcd-client\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785213 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/401d0883-625c-4bb6-80ba-8471f93c388b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785233 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/401d0883-625c-4bb6-80ba-8471f93c388b-encryption-config\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785315 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d755342a-48bb-467b-bc8a-5f18a8389ceb-machine-approver-tls\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785341 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-dir\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785361 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfkn8\" (UniqueName: \"kubernetes.io/projected/eb1393c2-0fc7-4150-a96e-d48ec1d45992-kube-api-access-pfkn8\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785380 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785401 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a8b9202a-16ae-4569-9b82-e7342d6b82c5-etcd-service-ca\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785422 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-oauth-config\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785438 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-trusted-ca-bundle\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785455 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2w2b\" (UniqueName: \"kubernetes.io/projected/44a13fe5-55be-4e50-b2c1-7d0ba939c613-kube-api-access-b2w2b\") pod \"migrator-59844c95c7-2456d\" (UID: \"44a13fe5-55be-4e50-b2c1-7d0ba939c613\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785478 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c37f83cf-f85a-4872-8e9e-ed37253c753f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z97pw\" (UID: \"c37f83cf-f85a-4872-8e9e-ed37253c753f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785637 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c37f95a3-8ee6-4cd3-bfd5-79f688335b19-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qw297\" (UID: \"c37f95a3-8ee6-4cd3-bfd5-79f688335b19\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785664 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bf69a9c-5575-4c6f-8751-5ca9f3640863-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvfh6\" (UID: \"3bf69a9c-5575-4c6f-8751-5ca9f3640863\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785713 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4464ee82-fef7-46d6-8118-1e1d685e63f6-serving-cert\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785732 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785764 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-service-ca-bundle\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785809 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-client-ca\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785838 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-serving-cert\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785858 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/981e0fe0-3fa9-403a-a147-5d31881cf83f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785876 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vmrkj\" (UID: \"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785906 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4464ee82-fef7-46d6-8118-1e1d685e63f6-etcd-client\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785922 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-image-import-ca\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785939 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4464ee82-fef7-46d6-8118-1e1d685e63f6-encryption-config\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.785955 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc5jz\" (UniqueName: \"kubernetes.io/projected/4464ee82-fef7-46d6-8118-1e1d685e63f6-kube-api-access-jc5jz\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.786034 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.786457 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c37f95a3-8ee6-4cd3-bfd5-79f688335b19-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qw297\" (UID: \"c37f95a3-8ee6-4cd3-bfd5-79f688335b19\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.787313 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a8b9202a-16ae-4569-9b82-e7342d6b82c5-etcd-ca\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.788001 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.788197 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-config\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.788932 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.790243 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8abb0401-2b84-4ae7-abe1-6e151c6b1bb6-profile-collector-cert\") pod \"catalog-operator-68c6474976-v2b95\" (UID: \"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.790297 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-config\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.790318 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.790340 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8b9202a-16ae-4569-9b82-e7342d6b82c5-serving-cert\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.790539 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c37f95a3-8ee6-4cd3-bfd5-79f688335b19-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qw297\" (UID: \"c37f95a3-8ee6-4cd3-bfd5-79f688335b19\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.790619 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-serving-cert\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.791319 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.794368 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-oauth-serving-cert\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.796583 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-client-ca\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.804265 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-7jnjl"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.805146 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-config\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.805480 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.805885 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-oauth-config\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.808076 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-serving-cert\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.808587 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.810063 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-trusted-ca-bundle\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.811154 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-service-ca\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.811820 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.812635 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.815561 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9rq8l"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.816903 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.818408 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-26ghk"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.820266 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.822784 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9sd7w"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.824942 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.828319 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.829136 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lwzgh"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.830687 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.833256 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.835732 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.837540 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.838514 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.839941 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.843741 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.848576 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.858930 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.861948 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.863395 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.864920 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-44mh6"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.866465 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.868012 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-897kb"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.869502 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5qd2g"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.870631 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.871160 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.872613 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.873877 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dtw6g"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.875145 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.876768 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2dxwn"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.878272 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.879945 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.880771 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-8fckr"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.881910 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5qc5h"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.882022 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.883138 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvsh5"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.884337 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b9mj6"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.886850 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.886986 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.888180 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p9k66"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.889598 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.889727 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8n9cm"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.891549 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8fckr"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.893030 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p9k66"] Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.893942 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0da29a-a520-4aa8-ae7a-47a22fc9a81e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jfbh\" (UID: \"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.894914 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45725226-1710-4204-8055-ecea69f828ee-serving-cert\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.894951 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72rqw\" (UniqueName: \"kubernetes.io/projected/9d11e36b-e397-47e6-a3d1-93cf9367f83f-kube-api-access-72rqw\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.894772 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895036 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895074 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-config\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895093 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/54841e96-db7c-4118-bbfb-3d35bbf59a4d-metrics-tls\") pod \"dns-operator-744455d44c-2dxwn\" (UID: \"54841e96-db7c-4118-bbfb-3d35bbf59a4d\") " pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895109 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f0da29a-a520-4aa8-ae7a-47a22fc9a81e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jfbh\" (UID: \"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895132 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895155 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7eebe686-79b8-4742-bd0b-09db708418d8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895173 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/68e9fb5c-167b-4528-9c31-97cbdc41aecd-default-certificate\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895193 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phtw6\" (UniqueName: \"kubernetes.io/projected/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-kube-api-access-phtw6\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895216 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-cabundle\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895237 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895253 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-key\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895276 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/792dbe38-981d-4360-bf53-4b79947bc556-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-d4vgh\" (UID: \"792dbe38-981d-4360-bf53-4b79947bc556\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895296 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-dir\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895316 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfkn8\" (UniqueName: \"kubernetes.io/projected/eb1393c2-0fc7-4150-a96e-d48ec1d45992-kube-api-access-pfkn8\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895347 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c37f83cf-f85a-4872-8e9e-ed37253c753f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z97pw\" (UID: \"c37f83cf-f85a-4872-8e9e-ed37253c753f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895394 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4464ee82-fef7-46d6-8118-1e1d685e63f6-serving-cert\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895415 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bf69a9c-5575-4c6f-8751-5ca9f3640863-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvfh6\" (UID: \"3bf69a9c-5575-4c6f-8751-5ca9f3640863\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895437 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/981e0fe0-3fa9-403a-a147-5d31881cf83f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895572 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vmrkj\" (UID: \"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895622 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-image-import-ca\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895648 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4464ee82-fef7-46d6-8118-1e1d685e63f6-encryption-config\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895674 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8abb0401-2b84-4ae7-abe1-6e151c6b1bb6-profile-collector-cert\") pod \"catalog-operator-68c6474976-v2b95\" (UID: \"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895698 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8b9202a-16ae-4569-9b82-e7342d6b82c5-serving-cert\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895720 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a8b9202a-16ae-4569-9b82-e7342d6b82c5-etcd-ca\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895744 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cecd9316-6fa9-4f5f-bc1c-aa468860ea3d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wb9v2\" (UID: \"cecd9316-6fa9-4f5f-bc1c-aa468860ea3d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895792 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895890 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8b9202a-16ae-4569-9b82-e7342d6b82c5-config\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895920 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xjcb\" (UniqueName: \"kubernetes.io/projected/54841e96-db7c-4118-bbfb-3d35bbf59a4d-kube-api-access-2xjcb\") pod \"dns-operator-744455d44c-2dxwn\" (UID: \"54841e96-db7c-4118-bbfb-3d35bbf59a4d\") " pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895945 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-serving-cert\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895968 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8e9f6b9-0e46-4271-89c3-657726d9dbed-serving-cert\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.895993 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ssrd\" (UniqueName: \"kubernetes.io/projected/45725226-1710-4204-8055-ecea69f828ee-kube-api-access-4ssrd\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896017 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-policies\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896042 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpvmn\" (UniqueName: \"kubernetes.io/projected/401d0883-625c-4bb6-80ba-8471f93c388b-kube-api-access-mpvmn\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896068 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8abb0401-2b84-4ae7-abe1-6e151c6b1bb6-srv-cert\") pod \"catalog-operator-68c6474976-v2b95\" (UID: \"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896096 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkbtn\" (UniqueName: \"kubernetes.io/projected/68e9fb5c-167b-4528-9c31-97cbdc41aecd-kube-api-access-tkbtn\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896108 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-config\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896122 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a8e9f6b9-0e46-4271-89c3-657726d9dbed-trusted-ca\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896147 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sddkp\" (UniqueName: \"kubernetes.io/projected/792dbe38-981d-4360-bf53-4b79947bc556-kube-api-access-sddkp\") pod \"machine-config-controller-84d6567774-d4vgh\" (UID: \"792dbe38-981d-4360-bf53-4b79947bc556\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896175 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b0d7439-9f5b-4a4f-ab08-64d77a966845-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zrf6j\" (UID: \"9b0d7439-9f5b-4a4f-ab08-64d77a966845\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896203 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/401d0883-625c-4bb6-80ba-8471f93c388b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896236 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d-serving-cert\") pod \"openshift-config-operator-7777fb866f-vmrkj\" (UID: \"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896264 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/215543fc-99e1-47f3-9ffd-9adb4fd9c611-secret-volume\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896289 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bc36fc2-f4d0-4381-b766-43da3213c356-serving-cert\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896317 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/401d0883-625c-4bb6-80ba-8471f93c388b-audit-policies\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896344 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89w7l\" (UniqueName: \"kubernetes.io/projected/a8b9202a-16ae-4569-9b82-e7342d6b82c5-kube-api-access-89w7l\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896370 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc36fc2-f4d0-4381-b766-43da3213c356-config\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896398 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4464ee82-fef7-46d6-8118-1e1d685e63f6-audit-dir\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896426 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7eebe686-79b8-4742-bd0b-09db708418d8-trusted-ca\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896455 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz9j5\" (UniqueName: \"kubernetes.io/projected/d06b0170-f718-4285-a25b-b5c07868e3c2-kube-api-access-jz9j5\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896486 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-client-ca\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896510 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/215543fc-99e1-47f3-9ffd-9adb4fd9c611-config-volume\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896539 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mw6r9\" (UniqueName: \"kubernetes.io/projected/6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0-kube-api-access-mw6r9\") pod \"multus-admission-controller-857f4d67dd-5qc5h\" (UID: \"6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896564 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d755342a-48bb-467b-bc8a-5f18a8389ceb-config\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896590 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896617 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d06b0170-f718-4285-a25b-b5c07868e3c2-config\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896640 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a8b9202a-16ae-4569-9b82-e7342d6b82c5-etcd-client\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896737 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9302abd4-6cf6-485b-97a1-0c0805226f4c-images\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896791 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-audit\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896832 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/981e0fe0-3fa9-403a-a147-5d31881cf83f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896860 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n56qw\" (UniqueName: \"kubernetes.io/projected/c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d-kube-api-access-n56qw\") pod \"openshift-config-operator-7777fb866f-vmrkj\" (UID: \"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896887 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7c8065b-1f23-4125-9ef0-4caa50b70c4c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jrmhr\" (UID: \"a7c8065b-1f23-4125-9ef0-4caa50b70c4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896914 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68e9fb5c-167b-4528-9c31-97cbdc41aecd-metrics-certs\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896963 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.896989 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d06b0170-f718-4285-a25b-b5c07868e3c2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.897137 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/792dbe38-981d-4360-bf53-4b79947bc556-proxy-tls\") pod \"machine-config-controller-84d6567774-d4vgh\" (UID: \"792dbe38-981d-4360-bf53-4b79947bc556\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.897186 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bf69a9c-5575-4c6f-8751-5ca9f3640863-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvfh6\" (UID: \"3bf69a9c-5575-4c6f-8751-5ca9f3640863\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.897452 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d755342a-48bb-467b-bc8a-5f18a8389ceb-machine-approver-tls\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.897555 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/401d0883-625c-4bb6-80ba-8471f93c388b-etcd-client\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.897585 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/401d0883-625c-4bb6-80ba-8471f93c388b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.897629 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/401d0883-625c-4bb6-80ba-8471f93c388b-encryption-config\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.897662 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7mjc\" (UniqueName: \"kubernetes.io/projected/9302abd4-6cf6-485b-97a1-0c0805226f4c-kube-api-access-z7mjc\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.897852 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0da29a-a520-4aa8-ae7a-47a22fc9a81e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jfbh\" (UID: \"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.897889 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898084 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a8b9202a-16ae-4569-9b82-e7342d6b82c5-etcd-service-ca\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898116 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2w2b\" (UniqueName: \"kubernetes.io/projected/44a13fe5-55be-4e50-b2c1-7d0ba939c613-kube-api-access-b2w2b\") pod \"migrator-59844c95c7-2456d\" (UID: \"44a13fe5-55be-4e50-b2c1-7d0ba939c613\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898143 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898170 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/815ec724-0b09-4f07-8b6e-37144aaebae0-config\") pod \"kube-controller-manager-operator-78b949d7b-lszpf\" (UID: \"815ec724-0b09-4f07-8b6e-37144aaebae0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898196 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898222 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-service-ca-bundle\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898263 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4464ee82-fef7-46d6-8118-1e1d685e63f6-etcd-client\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898289 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc5jz\" (UniqueName: \"kubernetes.io/projected/4464ee82-fef7-46d6-8118-1e1d685e63f6-kube-api-access-jc5jz\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898313 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-config\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898340 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898367 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28996\" (UniqueName: \"kubernetes.io/projected/981e0fe0-3fa9-403a-a147-5d31881cf83f-kube-api-access-28996\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898391 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7c8065b-1f23-4125-9ef0-4caa50b70c4c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jrmhr\" (UID: \"a7c8065b-1f23-4125-9ef0-4caa50b70c4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898421 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcpkj\" (UniqueName: \"kubernetes.io/projected/3bf69a9c-5575-4c6f-8751-5ca9f3640863-kube-api-access-wcpkj\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvfh6\" (UID: \"3bf69a9c-5575-4c6f-8751-5ca9f3640863\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898448 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/815ec724-0b09-4f07-8b6e-37144aaebae0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lszpf\" (UID: \"815ec724-0b09-4f07-8b6e-37144aaebae0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898478 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/981e0fe0-3fa9-403a-a147-5d31881cf83f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898505 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47zpm\" (UniqueName: \"kubernetes.io/projected/8abb0401-2b84-4ae7-abe1-6e151c6b1bb6-kube-api-access-47zpm\") pod \"catalog-operator-68c6474976-v2b95\" (UID: \"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898531 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4t64\" (UniqueName: \"kubernetes.io/projected/d755342a-48bb-467b-bc8a-5f18a8389ceb-kube-api-access-r4t64\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898556 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tp52\" (UniqueName: \"kubernetes.io/projected/a8e9f6b9-0e46-4271-89c3-657726d9dbed-kube-api-access-4tp52\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898588 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-config\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898863 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/68e9fb5c-167b-4528-9c31-97cbdc41aecd-stats-auth\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898897 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d755342a-48bb-467b-bc8a-5f18a8389ceb-auth-proxy-config\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898926 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898956 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrkqj\" (UniqueName: \"kubernetes.io/projected/9bc36fc2-f4d0-4381-b766-43da3213c356-kube-api-access-vrkqj\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.898987 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4464ee82-fef7-46d6-8118-1e1d685e63f6-node-pullsecrets\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.899014 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.899040 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/68e9fb5c-167b-4528-9c31-97cbdc41aecd-service-ca-bundle\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.899066 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5qc5h\" (UID: \"6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.899793 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.899918 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900059 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7f7v\" (UniqueName: \"kubernetes.io/projected/cecd9316-6fa9-4f5f-bc1c-aa468860ea3d-kube-api-access-j7f7v\") pod \"cluster-samples-operator-665b6dd947-wb9v2\" (UID: \"cecd9316-6fa9-4f5f-bc1c-aa468860ea3d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900131 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900172 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qsgx\" (UniqueName: \"kubernetes.io/projected/7eebe686-79b8-4742-bd0b-09db708418d8-kube-api-access-5qsgx\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900252 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-audit\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900267 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8e9f6b9-0e46-4271-89c3-657726d9dbed-config\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900315 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/815ec724-0b09-4f07-8b6e-37144aaebae0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lszpf\" (UID: \"815ec724-0b09-4f07-8b6e-37144aaebae0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900342 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9302abd4-6cf6-485b-97a1-0c0805226f4c-proxy-tls\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900354 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-service-ca-bundle\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900370 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs9tk\" (UniqueName: \"kubernetes.io/projected/215543fc-99e1-47f3-9ffd-9adb4fd9c611-kube-api-access-rs9tk\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900397 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900625 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9302abd4-6cf6-485b-97a1-0c0805226f4c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900664 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b0d7439-9f5b-4a4f-ab08-64d77a966845-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zrf6j\" (UID: \"9b0d7439-9f5b-4a4f-ab08-64d77a966845\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900686 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900707 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/401d0883-625c-4bb6-80ba-8471f93c388b-serving-cert\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900726 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/401d0883-625c-4bb6-80ba-8471f93c388b-audit-dir\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900763 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d06b0170-f718-4285-a25b-b5c07868e3c2-images\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900622 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-vmrkj\" (UID: \"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900845 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxwqh\" (UniqueName: \"kubernetes.io/projected/c37f83cf-f85a-4872-8e9e-ed37253c753f-kube-api-access-zxwqh\") pod \"control-plane-machine-set-operator-78cbb6b69f-z97pw\" (UID: \"c37f83cf-f85a-4872-8e9e-ed37253c753f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900890 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900921 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7eebe686-79b8-4742-bd0b-09db708418d8-metrics-tls\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900948 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8clcm\" (UniqueName: \"kubernetes.io/projected/e32ccf22-28fb-4b8e-8e58-2789580d723c-kube-api-access-8clcm\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900974 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cddaff14-4c07-4e69-b4be-0a16cb8db61d-cert\") pod \"ingress-canary-8n9cm\" (UID: \"cddaff14-4c07-4e69-b4be-0a16cb8db61d\") " pod="openshift-ingress-canary/ingress-canary-8n9cm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900999 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7c8065b-1f23-4125-9ef0-4caa50b70c4c-config\") pod \"kube-apiserver-operator-766d6c64bb-jrmhr\" (UID: \"a7c8065b-1f23-4125-9ef0-4caa50b70c4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.901032 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7qf5\" (UniqueName: \"kubernetes.io/projected/9b0d7439-9f5b-4a4f-ab08-64d77a966845-kube-api-access-z7qf5\") pod \"openshift-apiserver-operator-796bbdcf4f-zrf6j\" (UID: \"9b0d7439-9f5b-4a4f-ab08-64d77a966845\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.901056 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjcx4\" (UniqueName: \"kubernetes.io/projected/dc7bae79-3278-46f0-91fd-9a49b4057200-kube-api-access-rjcx4\") pod \"downloads-7954f5f757-9rq8l\" (UID: \"dc7bae79-3278-46f0-91fd-9a49b4057200\") " pod="openshift-console/downloads-7954f5f757-9rq8l" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.901277 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z496p\" (UniqueName: \"kubernetes.io/projected/cddaff14-4c07-4e69-b4be-0a16cb8db61d-kube-api-access-z496p\") pod \"ingress-canary-8n9cm\" (UID: \"cddaff14-4c07-4e69-b4be-0a16cb8db61d\") " pod="openshift-ingress-canary/ingress-canary-8n9cm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.901706 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b0d7439-9f5b-4a4f-ab08-64d77a966845-config\") pod \"openshift-apiserver-operator-796bbdcf4f-zrf6j\" (UID: \"9b0d7439-9f5b-4a4f-ab08-64d77a966845\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.901708 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-image-import-ca\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.900662 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-etcd-serving-ca\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.902418 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d06b0170-f718-4285-a25b-b5c07868e3c2-images\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.902533 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.903103 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/401d0883-625c-4bb6-80ba-8471f93c388b-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.903272 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45725226-1710-4204-8055-ecea69f828ee-serving-cert\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.903292 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.903387 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-trusted-ca-bundle\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.903911 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.904078 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-dir\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.904157 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/4464ee82-fef7-46d6-8118-1e1d685e63f6-etcd-client\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.904242 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8e9f6b9-0e46-4271-89c3-657726d9dbed-config\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.904839 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/401d0883-625c-4bb6-80ba-8471f93c388b-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.905017 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/4464ee82-fef7-46d6-8118-1e1d685e63f6-encryption-config\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.905068 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/981e0fe0-3fa9-403a-a147-5d31881cf83f-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.905170 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d755342a-48bb-467b-bc8a-5f18a8389ceb-auth-proxy-config\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.905339 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/d755342a-48bb-467b-bc8a-5f18a8389ceb-machine-approver-tls\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.906134 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-config\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.906085 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/401d0883-625c-4bb6-80ba-8471f93c388b-audit-policies\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.906200 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.906388 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/401d0883-625c-4bb6-80ba-8471f93c388b-audit-dir\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.906613 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-policies\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.906996 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.907034 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/401d0883-625c-4bb6-80ba-8471f93c388b-etcd-client\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.907047 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/4464ee82-fef7-46d6-8118-1e1d685e63f6-audit-dir\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.907386 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7eebe686-79b8-4742-bd0b-09db708418d8-trusted-ca\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.907429 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d06b0170-f718-4285-a25b-b5c07868e3c2-config\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.907997 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b0d7439-9f5b-4a4f-ab08-64d77a966845-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-zrf6j\" (UID: \"9b0d7439-9f5b-4a4f-ab08-64d77a966845\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.908111 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4464ee82-fef7-46d6-8118-1e1d685e63f6-config\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.908317 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8b9202a-16ae-4569-9b82-e7342d6b82c5-serving-cert\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.908525 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d755342a-48bb-467b-bc8a-5f18a8389ceb-config\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.908608 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a8e9f6b9-0e46-4271-89c3-657726d9dbed-trusted-ca\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.908747 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/4464ee82-fef7-46d6-8118-1e1d685e63f6-node-pullsecrets\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.908805 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/401d0883-625c-4bb6-80ba-8471f93c388b-encryption-config\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.909002 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a8b9202a-16ae-4569-9b82-e7342d6b82c5-etcd-ca\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.910045 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.910157 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-serving-cert\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.910505 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d-serving-cert\") pod \"openshift-config-operator-7777fb866f-vmrkj\" (UID: \"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.912195 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4464ee82-fef7-46d6-8118-1e1d685e63f6-serving-cert\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.913016 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/cecd9316-6fa9-4f5f-bc1c-aa468860ea3d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-wb9v2\" (UID: \"cecd9316-6fa9-4f5f-bc1c-aa468860ea3d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.913659 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.913794 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8e9f6b9-0e46-4271-89c3-657726d9dbed-serving-cert\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.913836 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7eebe686-79b8-4742-bd0b-09db708418d8-metrics-tls\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.913908 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/401d0883-625c-4bb6-80ba-8471f93c388b-serving-cert\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.914059 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/d06b0170-f718-4285-a25b-b5c07868e3c2-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.914176 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/981e0fe0-3fa9-403a-a147-5d31881cf83f-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.915295 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.915326 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.915503 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.915637 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a8b9202a-16ae-4569-9b82-e7342d6b82c5-etcd-client\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.917137 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.929264 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.934854 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a8b9202a-16ae-4569-9b82-e7342d6b82c5-etcd-service-ca\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.948727 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.959808 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-client-ca\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.969957 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.988760 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 09 10:07:34 crc kubenswrapper[4923]: I1009 10:07:34.999416 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8b9202a-16ae-4569-9b82-e7342d6b82c5-config\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002054 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrkqj\" (UniqueName: \"kubernetes.io/projected/9bc36fc2-f4d0-4381-b766-43da3213c356-kube-api-access-vrkqj\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002108 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002140 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9302abd4-6cf6-485b-97a1-0c0805226f4c-proxy-tls\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002265 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/815ec724-0b09-4f07-8b6e-37144aaebae0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lszpf\" (UID: \"815ec724-0b09-4f07-8b6e-37144aaebae0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002299 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs9tk\" (UniqueName: \"kubernetes.io/projected/215543fc-99e1-47f3-9ffd-9adb4fd9c611-kube-api-access-rs9tk\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002317 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9302abd4-6cf6-485b-97a1-0c0805226f4c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002343 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8clcm\" (UniqueName: \"kubernetes.io/projected/e32ccf22-28fb-4b8e-8e58-2789580d723c-kube-api-access-8clcm\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002362 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cddaff14-4c07-4e69-b4be-0a16cb8db61d-cert\") pod \"ingress-canary-8n9cm\" (UID: \"cddaff14-4c07-4e69-b4be-0a16cb8db61d\") " pod="openshift-ingress-canary/ingress-canary-8n9cm" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002378 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7c8065b-1f23-4125-9ef0-4caa50b70c4c-config\") pod \"kube-apiserver-operator-766d6c64bb-jrmhr\" (UID: \"a7c8065b-1f23-4125-9ef0-4caa50b70c4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002403 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z496p\" (UniqueName: \"kubernetes.io/projected/cddaff14-4c07-4e69-b4be-0a16cb8db61d-kube-api-access-z496p\") pod \"ingress-canary-8n9cm\" (UID: \"cddaff14-4c07-4e69-b4be-0a16cb8db61d\") " pod="openshift-ingress-canary/ingress-canary-8n9cm" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002427 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0da29a-a520-4aa8-ae7a-47a22fc9a81e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jfbh\" (UID: \"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002446 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72rqw\" (UniqueName: \"kubernetes.io/projected/9d11e36b-e397-47e6-a3d1-93cf9367f83f-kube-api-access-72rqw\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002475 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f0da29a-a520-4aa8-ae7a-47a22fc9a81e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jfbh\" (UID: \"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002506 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-cabundle\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002524 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-key\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.002543 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/792dbe38-981d-4360-bf53-4b79947bc556-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-d4vgh\" (UID: \"792dbe38-981d-4360-bf53-4b79947bc556\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.007159 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9302abd4-6cf6-485b-97a1-0c0805226f4c-auth-proxy-config\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.008479 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.013513 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/792dbe38-981d-4360-bf53-4b79947bc556-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-d4vgh\" (UID: \"792dbe38-981d-4360-bf53-4b79947bc556\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.013698 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sddkp\" (UniqueName: \"kubernetes.io/projected/792dbe38-981d-4360-bf53-4b79947bc556-kube-api-access-sddkp\") pod \"machine-config-controller-84d6567774-d4vgh\" (UID: \"792dbe38-981d-4360-bf53-4b79947bc556\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.013814 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/215543fc-99e1-47f3-9ffd-9adb4fd9c611-secret-volume\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.013844 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bc36fc2-f4d0-4381-b766-43da3213c356-serving-cert\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.013885 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc36fc2-f4d0-4381-b766-43da3213c356-config\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.013920 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/215543fc-99e1-47f3-9ffd-9adb4fd9c611-config-volume\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.013961 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9302abd4-6cf6-485b-97a1-0c0805226f4c-images\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.014011 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7c8065b-1f23-4125-9ef0-4caa50b70c4c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jrmhr\" (UID: \"a7c8065b-1f23-4125-9ef0-4caa50b70c4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.014065 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/792dbe38-981d-4360-bf53-4b79947bc556-proxy-tls\") pod \"machine-config-controller-84d6567774-d4vgh\" (UID: \"792dbe38-981d-4360-bf53-4b79947bc556\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.014100 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7mjc\" (UniqueName: \"kubernetes.io/projected/9302abd4-6cf6-485b-97a1-0c0805226f4c-kube-api-access-z7mjc\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.014119 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0da29a-a520-4aa8-ae7a-47a22fc9a81e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jfbh\" (UID: \"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.014155 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.014175 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/815ec724-0b09-4f07-8b6e-37144aaebae0-config\") pod \"kube-controller-manager-operator-78b949d7b-lszpf\" (UID: \"815ec724-0b09-4f07-8b6e-37144aaebae0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.014227 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7c8065b-1f23-4125-9ef0-4caa50b70c4c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jrmhr\" (UID: \"a7c8065b-1f23-4125-9ef0-4caa50b70c4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.014258 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/815ec724-0b09-4f07-8b6e-37144aaebae0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lszpf\" (UID: \"815ec724-0b09-4f07-8b6e-37144aaebae0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.018955 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/54841e96-db7c-4118-bbfb-3d35bbf59a4d-metrics-tls\") pod \"dns-operator-744455d44c-2dxwn\" (UID: \"54841e96-db7c-4118-bbfb-3d35bbf59a4d\") " pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.029155 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.049366 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.068289 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.088513 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.109346 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.127980 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.149134 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.151682 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bf69a9c-5575-4c6f-8751-5ca9f3640863-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvfh6\" (UID: \"3bf69a9c-5575-4c6f-8751-5ca9f3640863\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.167617 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.188863 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.209951 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.223826 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bf69a9c-5575-4c6f-8751-5ca9f3640863-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvfh6\" (UID: \"3bf69a9c-5575-4c6f-8751-5ca9f3640863\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.228349 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.232880 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/68e9fb5c-167b-4528-9c31-97cbdc41aecd-default-certificate\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.248856 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.257707 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/68e9fb5c-167b-4528-9c31-97cbdc41aecd-stats-auth\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.268636 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.273021 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68e9fb5c-167b-4528-9c31-97cbdc41aecd-metrics-certs\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.288850 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.309934 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.319270 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/68e9fb5c-167b-4528-9c31-97cbdc41aecd-service-ca-bundle\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.328177 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.348959 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.368700 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.375000 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c37f83cf-f85a-4872-8e9e-ed37253c753f-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-z97pw\" (UID: \"c37f83cf-f85a-4872-8e9e-ed37253c753f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.388794 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.409302 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.418901 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-5qc5h\" (UID: \"6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.429264 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.450489 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.468386 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.481631 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/8abb0401-2b84-4ae7-abe1-6e151c6b1bb6-srv-cert\") pod \"catalog-operator-68c6474976-v2b95\" (UID: \"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.489401 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.497711 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/215543fc-99e1-47f3-9ffd-9adb4fd9c611-secret-volume\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.499095 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/8abb0401-2b84-4ae7-abe1-6e151c6b1bb6-profile-collector-cert\") pod \"catalog-operator-68c6474976-v2b95\" (UID: \"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.509134 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.528859 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.569045 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.579349 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/792dbe38-981d-4360-bf53-4b79947bc556-proxy-tls\") pod \"machine-config-controller-84d6567774-d4vgh\" (UID: \"792dbe38-981d-4360-bf53-4b79947bc556\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.590164 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.609706 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.629318 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.650517 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.658054 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/815ec724-0b09-4f07-8b6e-37144aaebae0-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-lszpf\" (UID: \"815ec724-0b09-4f07-8b6e-37144aaebae0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.669525 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.676172 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/815ec724-0b09-4f07-8b6e-37144aaebae0-config\") pod \"kube-controller-manager-operator-78b949d7b-lszpf\" (UID: \"815ec724-0b09-4f07-8b6e-37144aaebae0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.689167 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.708746 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.727547 4923 request.go:700] Waited for 1.012169607s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-scheduler-operator/secrets?fieldSelector=metadata.name%3Dkube-scheduler-operator-serving-cert&limit=500&resourceVersion=0 Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.729146 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.738252 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0f0da29a-a520-4aa8-ae7a-47a22fc9a81e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jfbh\" (UID: \"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.749582 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.753457 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0f0da29a-a520-4aa8-ae7a-47a22fc9a81e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jfbh\" (UID: \"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.769182 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.788432 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.811335 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.829028 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.848625 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.858237 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a7c8065b-1f23-4125-9ef0-4caa50b70c4c-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-jrmhr\" (UID: \"a7c8065b-1f23-4125-9ef0-4caa50b70c4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.868208 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.874040 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a7c8065b-1f23-4125-9ef0-4caa50b70c4c-config\") pod \"kube-apiserver-operator-766d6c64bb-jrmhr\" (UID: \"a7c8065b-1f23-4125-9ef0-4caa50b70c4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.888879 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.908992 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.915826 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9302abd4-6cf6-485b-97a1-0c0805226f4c-images\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.929344 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.949512 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.956730 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9302abd4-6cf6-485b-97a1-0c0805226f4c-proxy-tls\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.982359 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.986194 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:35 crc kubenswrapper[4923]: I1009 10:07:35.989063 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.002644 4923 secret.go:188] Couldn't get secret openshift-marketplace/marketplace-operator-metrics: failed to sync secret cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.002778 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-operator-metrics podName:9d11e36b-e397-47e6-a3d1-93cf9367f83f nodeName:}" failed. No retries permitted until 2025-10-09 10:07:36.502728648 +0000 UTC m=+142.570910404 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-operator-metrics" (UniqueName: "kubernetes.io/secret/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-operator-metrics") pod "marketplace-operator-79b997595-897kb" (UID: "9d11e36b-e397-47e6-a3d1-93cf9367f83f") : failed to sync secret cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.002662 4923 secret.go:188] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.002853 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cddaff14-4c07-4e69-b4be-0a16cb8db61d-cert podName:cddaff14-4c07-4e69-b4be-0a16cb8db61d nodeName:}" failed. No retries permitted until 2025-10-09 10:07:36.502837882 +0000 UTC m=+142.571019638 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/cddaff14-4c07-4e69-b4be-0a16cb8db61d-cert") pod "ingress-canary-8n9cm" (UID: "cddaff14-4c07-4e69-b4be-0a16cb8db61d") : failed to sync secret cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.006981 4923 secret.go:188] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.007043 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-key podName:e32ccf22-28fb-4b8e-8e58-2789580d723c nodeName:}" failed. No retries permitted until 2025-10-09 10:07:36.507029274 +0000 UTC m=+142.575211030 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-key") pod "service-ca-9c57cc56f-b9mj6" (UID: "e32ccf22-28fb-4b8e-8e58-2789580d723c") : failed to sync secret cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.007137 4923 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.007249 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-cabundle podName:e32ccf22-28fb-4b8e-8e58-2789580d723c nodeName:}" failed. No retries permitted until 2025-10-09 10:07:36.507221249 +0000 UTC m=+142.575403175 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-cabundle") pod "service-ca-9c57cc56f-b9mj6" (UID: "e32ccf22-28fb-4b8e-8e58-2789580d723c") : failed to sync configmap cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.008963 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.014860 4923 secret.go:188] Couldn't get secret openshift-service-ca-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.014939 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9bc36fc2-f4d0-4381-b766-43da3213c356-serving-cert podName:9bc36fc2-f4d0-4381-b766-43da3213c356 nodeName:}" failed. No retries permitted until 2025-10-09 10:07:36.514919158 +0000 UTC m=+142.583101084 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/9bc36fc2-f4d0-4381-b766-43da3213c356-serving-cert") pod "service-ca-operator-777779d784-2zlvc" (UID: "9bc36fc2-f4d0-4381-b766-43da3213c356") : failed to sync secret cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.014994 4923 configmap.go:193] Couldn't get configMap openshift-service-ca-operator/service-ca-operator-config: failed to sync configmap cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.015024 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/9bc36fc2-f4d0-4381-b766-43da3213c356-config podName:9bc36fc2-f4d0-4381-b766-43da3213c356 nodeName:}" failed. No retries permitted until 2025-10-09 10:07:36.51501501 +0000 UTC m=+142.583196966 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/9bc36fc2-f4d0-4381-b766-43da3213c356-config") pod "service-ca-operator-777779d784-2zlvc" (UID: "9bc36fc2-f4d0-4381-b766-43da3213c356") : failed to sync configmap cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.014878 4923 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: E1009 10:07:36.015316 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/215543fc-99e1-47f3-9ffd-9adb4fd9c611-config-volume podName:215543fc-99e1-47f3-9ffd-9adb4fd9c611 nodeName:}" failed. No retries permitted until 2025-10-09 10:07:36.515295688 +0000 UTC m=+142.583477444 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/215543fc-99e1-47f3-9ffd-9adb4fd9c611-config-volume") pod "collect-profiles-29333400-d47m2" (UID: "215543fc-99e1-47f3-9ffd-9adb4fd9c611") : failed to sync configmap cache: timed out waiting for the condition Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.027894 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.056733 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.068774 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.089665 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.108782 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.128852 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.149451 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.169598 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.190016 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.209684 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.229515 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.249063 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.268671 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.289741 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.324884 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh754\" (UniqueName: \"kubernetes.io/projected/a8c487fa-3436-4cc9-83b3-de49256e97c9-kube-api-access-bh754\") pod \"console-f9d7485db-66f2z\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.329213 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.349206 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.369665 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.389031 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.424818 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dz5nz\" (UniqueName: \"kubernetes.io/projected/c37f95a3-8ee6-4cd3-bfd5-79f688335b19-kube-api-access-dz5nz\") pod \"openshift-controller-manager-operator-756b6f6bc6-qw297\" (UID: \"c37f95a3-8ee6-4cd3-bfd5-79f688335b19\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.426580 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.445133 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx26v\" (UniqueName: \"kubernetes.io/projected/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-kube-api-access-nx26v\") pod \"controller-manager-879f6c89f-tnbft\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.448816 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.469397 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.488916 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.510045 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.530094 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.538340 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-cabundle\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.538372 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-key\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.538524 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bc36fc2-f4d0-4381-b766-43da3213c356-serving-cert\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.538555 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc36fc2-f4d0-4381-b766-43da3213c356-config\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.538596 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/215543fc-99e1-47f3-9ffd-9adb4fd9c611-config-volume\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.538745 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.538850 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cddaff14-4c07-4e69-b4be-0a16cb8db61d-cert\") pod \"ingress-canary-8n9cm\" (UID: \"cddaff14-4c07-4e69-b4be-0a16cb8db61d\") " pod="openshift-ingress-canary/ingress-canary-8n9cm" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.541375 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9bc36fc2-f4d0-4381-b766-43da3213c356-config\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.542309 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-cabundle\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.542339 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/215543fc-99e1-47f3-9ffd-9adb4fd9c611-config-volume\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.542932 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cddaff14-4c07-4e69-b4be-0a16cb8db61d-cert\") pod \"ingress-canary-8n9cm\" (UID: \"cddaff14-4c07-4e69-b4be-0a16cb8db61d\") " pod="openshift-ingress-canary/ingress-canary-8n9cm" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.544163 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e32ccf22-28fb-4b8e-8e58-2789580d723c-signing-key\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.545108 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.545564 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9bc36fc2-f4d0-4381-b766-43da3213c356-serving-cert\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.550172 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.569342 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.589700 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.610209 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.610271 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297"] Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.628570 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.650605 4923 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.669397 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.670070 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-66f2z"] Oct 09 10:07:36 crc kubenswrapper[4923]: W1009 10:07:36.677815 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda8c487fa_3436_4cc9_83b3_de49256e97c9.slice/crio-cb8e7efd2a6f964b7d205e396df38dce86f7f139beb9137f97b52b007f3113b8 WatchSource:0}: Error finding container cb8e7efd2a6f964b7d205e396df38dce86f7f139beb9137f97b52b007f3113b8: Status 404 returned error can't find the container with id cb8e7efd2a6f964b7d205e396df38dce86f7f139beb9137f97b52b007f3113b8 Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.689103 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.711117 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.724527 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n56qw\" (UniqueName: \"kubernetes.io/projected/c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d-kube-api-access-n56qw\") pod \"openshift-config-operator-7777fb866f-vmrkj\" (UID: \"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.745876 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qsgx\" (UniqueName: \"kubernetes.io/projected/7eebe686-79b8-4742-bd0b-09db708418d8-kube-api-access-5qsgx\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.747885 4923 request.go:700] Waited for 1.847255916s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-operator/serviceaccounts/ingress-operator/token Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.769957 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7eebe686-79b8-4742-bd0b-09db708418d8-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sgdx2\" (UID: \"7eebe686-79b8-4742-bd0b-09db708418d8\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.783738 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7f7v\" (UniqueName: \"kubernetes.io/projected/cecd9316-6fa9-4f5f-bc1c-aa468860ea3d-kube-api-access-j7f7v\") pod \"cluster-samples-operator-665b6dd947-wb9v2\" (UID: \"cecd9316-6fa9-4f5f-bc1c-aa468860ea3d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.817452 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phtw6\" (UniqueName: \"kubernetes.io/projected/681bc3f8-88f8-4d02-ba8c-d74ee2e93058-kube-api-access-phtw6\") pod \"authentication-operator-69f744f599-26ghk\" (UID: \"681bc3f8-88f8-4d02-ba8c-d74ee2e93058\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.823671 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz9j5\" (UniqueName: \"kubernetes.io/projected/d06b0170-f718-4285-a25b-b5c07868e3c2-kube-api-access-jz9j5\") pod \"machine-api-operator-5694c8668f-9sd7w\" (UID: \"d06b0170-f718-4285-a25b-b5c07868e3c2\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.852633 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfkn8\" (UniqueName: \"kubernetes.io/projected/eb1393c2-0fc7-4150-a96e-d48ec1d45992-kube-api-access-pfkn8\") pod \"oauth-openshift-558db77b4-lwzgh\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.868179 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2w2b\" (UniqueName: \"kubernetes.io/projected/44a13fe5-55be-4e50-b2c1-7d0ba939c613-kube-api-access-b2w2b\") pod \"migrator-59844c95c7-2456d\" (UID: \"44a13fe5-55be-4e50-b2c1-7d0ba939c613\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.870663 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.881152 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.886731 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc5jz\" (UniqueName: \"kubernetes.io/projected/4464ee82-fef7-46d6-8118-1e1d685e63f6-kube-api-access-jc5jz\") pod \"apiserver-76f77b778f-dtw6g\" (UID: \"4464ee82-fef7-46d6-8118-1e1d685e63f6\") " pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.909927 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxwqh\" (UniqueName: \"kubernetes.io/projected/c37f83cf-f85a-4872-8e9e-ed37253c753f-kube-api-access-zxwqh\") pod \"control-plane-machine-set-operator-78cbb6b69f-z97pw\" (UID: \"c37f83cf-f85a-4872-8e9e-ed37253c753f\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.917401 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.922102 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tnbft"] Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.925237 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ssrd\" (UniqueName: \"kubernetes.io/projected/45725226-1710-4204-8055-ecea69f828ee-kube-api-access-4ssrd\") pod \"route-controller-manager-6576b87f9c-d8h64\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:36 crc kubenswrapper[4923]: W1009 10:07:36.935904 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7df9ce28_8006_4c4f_9d21_4c0ef83d2f44.slice/crio-5311fe0dca51dc4392f58a2b7fa77e4b8a70f6e55e9de94dadbd9f9660ee0abb WatchSource:0}: Error finding container 5311fe0dca51dc4392f58a2b7fa77e4b8a70f6e55e9de94dadbd9f9660ee0abb: Status 404 returned error can't find the container with id 5311fe0dca51dc4392f58a2b7fa77e4b8a70f6e55e9de94dadbd9f9660ee0abb Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.938872 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.945383 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tp52\" (UniqueName: \"kubernetes.io/projected/a8e9f6b9-0e46-4271-89c3-657726d9dbed-kube-api-access-4tp52\") pod \"console-operator-58897d9998-5qd2g\" (UID: \"a8e9f6b9-0e46-4271-89c3-657726d9dbed\") " pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.946141 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.954973 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.969305 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.973342 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcpkj\" (UniqueName: \"kubernetes.io/projected/3bf69a9c-5575-4c6f-8751-5ca9f3640863-kube-api-access-wcpkj\") pod \"kube-storage-version-migrator-operator-b67b599dd-rvfh6\" (UID: \"3bf69a9c-5575-4c6f-8751-5ca9f3640863\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:36 crc kubenswrapper[4923]: I1009 10:07:36.995425 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.009027 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4t64\" (UniqueName: \"kubernetes.io/projected/d755342a-48bb-467b-bc8a-5f18a8389ceb-kube-api-access-r4t64\") pod \"machine-approver-56656f9798-gvl7r\" (UID: \"d755342a-48bb-467b-bc8a-5f18a8389ceb\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.016245 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/981e0fe0-3fa9-403a-a147-5d31881cf83f-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.029809 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47zpm\" (UniqueName: \"kubernetes.io/projected/8abb0401-2b84-4ae7-abe1-6e151c6b1bb6-kube-api-access-47zpm\") pod \"catalog-operator-68c6474976-v2b95\" (UID: \"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.045043 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.059270 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mw6r9\" (UniqueName: \"kubernetes.io/projected/6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0-kube-api-access-mw6r9\") pod \"multus-admission-controller-857f4d67dd-5qc5h\" (UID: \"6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.061186 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.069820 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.078187 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.079369 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpvmn\" (UniqueName: \"kubernetes.io/projected/401d0883-625c-4bb6-80ba-8471f93c388b-kube-api-access-mpvmn\") pod \"apiserver-7bbb656c7d-gv7qk\" (UID: \"401d0883-625c-4bb6-80ba-8471f93c388b\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.086224 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.113640 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkbtn\" (UniqueName: \"kubernetes.io/projected/68e9fb5c-167b-4528-9c31-97cbdc41aecd-kube-api-access-tkbtn\") pod \"router-default-5444994796-85h6b\" (UID: \"68e9fb5c-167b-4528-9c31-97cbdc41aecd\") " pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.115469 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xjcb\" (UniqueName: \"kubernetes.io/projected/54841e96-db7c-4118-bbfb-3d35bbf59a4d-kube-api-access-2xjcb\") pod \"dns-operator-744455d44c-2dxwn\" (UID: \"54841e96-db7c-4118-bbfb-3d35bbf59a4d\") " pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.136646 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjcx4\" (UniqueName: \"kubernetes.io/projected/dc7bae79-3278-46f0-91fd-9a49b4057200-kube-api-access-rjcx4\") pod \"downloads-7954f5f757-9rq8l\" (UID: \"dc7bae79-3278-46f0-91fd-9a49b4057200\") " pod="openshift-console/downloads-7954f5f757-9rq8l" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.141377 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9sd7w"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.151670 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7qf5\" (UniqueName: \"kubernetes.io/projected/9b0d7439-9f5b-4a4f-ab08-64d77a966845-kube-api-access-z7qf5\") pod \"openshift-apiserver-operator-796bbdcf4f-zrf6j\" (UID: \"9b0d7439-9f5b-4a4f-ab08-64d77a966845\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:37 crc kubenswrapper[4923]: W1009 10:07:37.157186 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd06b0170_f718_4285_a25b_b5c07868e3c2.slice/crio-9c35071005fca1f635755de986d0e70929fe48ef9ae795fc4fdac768a1dee6bd WatchSource:0}: Error finding container 9c35071005fca1f635755de986d0e70929fe48ef9ae795fc4fdac768a1dee6bd: Status 404 returned error can't find the container with id 9c35071005fca1f635755de986d0e70929fe48ef9ae795fc4fdac768a1dee6bd Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.161718 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.166330 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.172953 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28996\" (UniqueName: \"kubernetes.io/projected/981e0fe0-3fa9-403a-a147-5d31881cf83f-kube-api-access-28996\") pod \"cluster-image-registry-operator-dc59b4c8b-h2spm\" (UID: \"981e0fe0-3fa9-403a-a147-5d31881cf83f\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.185491 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89w7l\" (UniqueName: \"kubernetes.io/projected/a8b9202a-16ae-4569-9b82-e7342d6b82c5-kube-api-access-89w7l\") pod \"etcd-operator-b45778765-44mh6\" (UID: \"a8b9202a-16ae-4569-9b82-e7342d6b82c5\") " pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.207716 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.208597 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrkqj\" (UniqueName: \"kubernetes.io/projected/9bc36fc2-f4d0-4381-b766-43da3213c356-kube-api-access-vrkqj\") pod \"service-ca-operator-777779d784-2zlvc\" (UID: \"9bc36fc2-f4d0-4381-b766-43da3213c356\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.224960 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.232228 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.233679 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/815ec724-0b09-4f07-8b6e-37144aaebae0-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-lszpf\" (UID: \"815ec724-0b09-4f07-8b6e-37144aaebae0\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.252317 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8clcm\" (UniqueName: \"kubernetes.io/projected/e32ccf22-28fb-4b8e-8e58-2789580d723c-kube-api-access-8clcm\") pod \"service-ca-9c57cc56f-b9mj6\" (UID: \"e32ccf22-28fb-4b8e-8e58-2789580d723c\") " pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.258299 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" event={"ID":"c37f95a3-8ee6-4cd3-bfd5-79f688335b19","Type":"ContainerStarted","Data":"65d59ddd65285901a6ff51ea670bc291787b9342827a338603182f5cbca27621"} Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.258346 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" event={"ID":"c37f95a3-8ee6-4cd3-bfd5-79f688335b19","Type":"ContainerStarted","Data":"3a309fb5a2145e5e74e71bbf6c863aeec8474c2600e64add16d6eccbcb68c29d"} Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.276098 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.277324 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z496p\" (UniqueName: \"kubernetes.io/projected/cddaff14-4c07-4e69-b4be-0a16cb8db61d-kube-api-access-z496p\") pod \"ingress-canary-8n9cm\" (UID: \"cddaff14-4c07-4e69-b4be-0a16cb8db61d\") " pod="openshift-ingress-canary/ingress-canary-8n9cm" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.279224 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" event={"ID":"d06b0170-f718-4285-a25b-b5c07868e3c2","Type":"ContainerStarted","Data":"9c35071005fca1f635755de986d0e70929fe48ef9ae795fc4fdac768a1dee6bd"} Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.280908 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" event={"ID":"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44","Type":"ContainerStarted","Data":"0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f"} Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.280934 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" event={"ID":"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44","Type":"ContainerStarted","Data":"5311fe0dca51dc4392f58a2b7fa77e4b8a70f6e55e9de94dadbd9f9660ee0abb"} Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.282034 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.282807 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.287585 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-66f2z" event={"ID":"a8c487fa-3436-4cc9-83b3-de49256e97c9","Type":"ContainerStarted","Data":"564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6"} Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.287842 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-66f2z" event={"ID":"a8c487fa-3436-4cc9-83b3-de49256e97c9","Type":"ContainerStarted","Data":"cb8e7efd2a6f964b7d205e396df38dce86f7f139beb9137f97b52b007f3113b8"} Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.288975 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.296703 4923 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-tnbft container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.296783 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" podUID="7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.305491 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0f0da29a-a520-4aa8-ae7a-47a22fc9a81e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6jfbh\" (UID: \"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.306121 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" event={"ID":"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d","Type":"ContainerStarted","Data":"703598aae7d76af722c5ed3e86d83b65bb29da951f239edd129c0654f55e710e"} Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.312126 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72rqw\" (UniqueName: \"kubernetes.io/projected/9d11e36b-e397-47e6-a3d1-93cf9367f83f-kube-api-access-72rqw\") pod \"marketplace-operator-79b997595-897kb\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.341710 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.346134 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs9tk\" (UniqueName: \"kubernetes.io/projected/215543fc-99e1-47f3-9ffd-9adb4fd9c611-kube-api-access-rs9tk\") pod \"collect-profiles-29333400-d47m2\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.350687 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.358720 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7c8065b-1f23-4125-9ef0-4caa50b70c4c-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-jrmhr\" (UID: \"a7c8065b-1f23-4125-9ef0-4caa50b70c4c\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.360801 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.376856 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sddkp\" (UniqueName: \"kubernetes.io/projected/792dbe38-981d-4360-bf53-4b79947bc556-kube-api-access-sddkp\") pod \"machine-config-controller-84d6567774-d4vgh\" (UID: \"792dbe38-981d-4360-bf53-4b79947bc556\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.389488 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7mjc\" (UniqueName: \"kubernetes.io/projected/9302abd4-6cf6-485b-97a1-0c0805226f4c-kube-api-access-z7mjc\") pod \"machine-config-operator-74547568cd-ndvn4\" (UID: \"9302abd4-6cf6-485b-97a1-0c0805226f4c\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.394608 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.404350 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.404435 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.419010 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.423509 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lwzgh"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.435066 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-dtw6g"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.435472 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-9rq8l" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.440037 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.448615 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.457660 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.467195 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469032 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cc177fc5-d31c-4488-ac3f-786acf59fd09-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5qtv8\" (UID: \"cc177fc5-d31c-4488-ac3f-786acf59fd09\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469105 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f4cb0dc-63c3-43d4-948a-4a355d083c2c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b9kd6\" (UID: \"4f4cb0dc-63c3-43d4-948a-4a355d083c2c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469153 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/02a21759-421a-429b-ac1b-c43cd7c236f0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469205 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cc177fc5-d31c-4488-ac3f-786acf59fd09-srv-cert\") pod \"olm-operator-6b444d44fb-5qtv8\" (UID: \"cc177fc5-d31c-4488-ac3f-786acf59fd09\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469249 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-bound-sa-token\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469274 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-trusted-ca\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469306 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/02a21759-421a-429b-ac1b-c43cd7c236f0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469327 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-tls\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469348 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7ngc\" (UniqueName: \"kubernetes.io/projected/4f4cb0dc-63c3-43d4-948a-4a355d083c2c-kube-api-access-z7ngc\") pod \"package-server-manager-789f6589d5-b9kd6\" (UID: \"4f4cb0dc-63c3-43d4-948a-4a355d083c2c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469424 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-certificates\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469451 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5pxp\" (UniqueName: \"kubernetes.io/projected/cc177fc5-d31c-4488-ac3f-786acf59fd09-kube-api-access-t5pxp\") pod \"olm-operator-6b444d44fb-5qtv8\" (UID: \"cc177fc5-d31c-4488-ac3f-786acf59fd09\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469480 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vh6l\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-kube-api-access-8vh6l\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469529 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.469634 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-26ghk"] Oct 09 10:07:37 crc kubenswrapper[4923]: E1009 10:07:37.469934 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:37.969918816 +0000 UTC m=+144.038100752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.488956 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.506644 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-8n9cm" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.513655 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.570417 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:37 crc kubenswrapper[4923]: E1009 10:07:37.571018 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.070957973 +0000 UTC m=+144.139139729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.571537 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cc177fc5-d31c-4488-ac3f-786acf59fd09-srv-cert\") pod \"olm-operator-6b444d44fb-5qtv8\" (UID: \"cc177fc5-d31c-4488-ac3f-786acf59fd09\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.571630 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whwqb\" (UniqueName: \"kubernetes.io/projected/7aba3bac-1dab-4f36-9ccf-2496c799d54d-kube-api-access-whwqb\") pod \"dns-default-8fckr\" (UID: \"7aba3bac-1dab-4f36-9ccf-2496c799d54d\") " pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.571655 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4c6e281d-d5ff-4554-874e-a17cacea5b5a-certs\") pod \"machine-config-server-7jnjl\" (UID: \"4c6e281d-d5ff-4554-874e-a17cacea5b5a\") " pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.571693 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7aba3bac-1dab-4f36-9ccf-2496c799d54d-metrics-tls\") pod \"dns-default-8fckr\" (UID: \"7aba3bac-1dab-4f36-9ccf-2496c799d54d\") " pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.571711 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96098499-0c0f-4650-8f9f-86094e45f19a-apiservice-cert\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.571783 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2xwm\" (UniqueName: \"kubernetes.io/projected/4c6e281d-d5ff-4554-874e-a17cacea5b5a-kube-api-access-l2xwm\") pod \"machine-config-server-7jnjl\" (UID: \"4c6e281d-d5ff-4554-874e-a17cacea5b5a\") " pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.575358 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-bound-sa-token\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.575814 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-trusted-ca\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.576501 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/02a21759-421a-429b-ac1b-c43cd7c236f0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.576566 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-registration-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.576689 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-tls\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.576718 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7ngc\" (UniqueName: \"kubernetes.io/projected/4f4cb0dc-63c3-43d4-948a-4a355d083c2c-kube-api-access-z7ngc\") pod \"package-server-manager-789f6589d5-b9kd6\" (UID: \"4f4cb0dc-63c3-43d4-948a-4a355d083c2c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.578064 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-trusted-ca\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.592390 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/cc177fc5-d31c-4488-ac3f-786acf59fd09-srv-cert\") pod \"olm-operator-6b444d44fb-5qtv8\" (UID: \"cc177fc5-d31c-4488-ac3f-786acf59fd09\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.595238 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/02a21759-421a-429b-ac1b-c43cd7c236f0-ca-trust-extracted\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.603848 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5pxp\" (UniqueName: \"kubernetes.io/projected/cc177fc5-d31c-4488-ac3f-786acf59fd09-kube-api-access-t5pxp\") pod \"olm-operator-6b444d44fb-5qtv8\" (UID: \"cc177fc5-d31c-4488-ac3f-786acf59fd09\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.603980 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4c6e281d-d5ff-4554-874e-a17cacea5b5a-node-bootstrap-token\") pod \"machine-config-server-7jnjl\" (UID: \"4c6e281d-d5ff-4554-874e-a17cacea5b5a\") " pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.604060 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-certificates\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.605020 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vh6l\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-kube-api-access-8vh6l\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.605116 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-csi-data-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.605200 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwdrf\" (UniqueName: \"kubernetes.io/projected/eb5ecfa1-07aa-48ca-8727-a28a935fd937-kube-api-access-wwdrf\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.605446 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.605519 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cc177fc5-d31c-4488-ac3f-786acf59fd09-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5qtv8\" (UID: \"cc177fc5-d31c-4488-ac3f-786acf59fd09\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.605651 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blp9l\" (UniqueName: \"kubernetes.io/projected/96098499-0c0f-4650-8f9f-86094e45f19a-kube-api-access-blp9l\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.606059 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/96098499-0c0f-4650-8f9f-86094e45f19a-tmpfs\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.606221 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-socket-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.606369 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f4cb0dc-63c3-43d4-948a-4a355d083c2c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b9kd6\" (UID: \"4f4cb0dc-63c3-43d4-948a-4a355d083c2c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.610446 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-certificates\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.615798 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-tls\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.627699 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/cc177fc5-d31c-4488-ac3f-786acf59fd09-profile-collector-cert\") pod \"olm-operator-6b444d44fb-5qtv8\" (UID: \"cc177fc5-d31c-4488-ac3f-786acf59fd09\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.642027 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-bound-sa-token\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: E1009 10:07:37.644542 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.144503759 +0000 UTC m=+144.212685505 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.644958 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-plugins-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.646793 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-mountpoint-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.647330 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7aba3bac-1dab-4f36-9ccf-2496c799d54d-config-volume\") pod \"dns-default-8fckr\" (UID: \"7aba3bac-1dab-4f36-9ccf-2496c799d54d\") " pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.647431 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96098499-0c0f-4650-8f9f-86094e45f19a-webhook-cert\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.647479 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/02a21759-421a-429b-ac1b-c43cd7c236f0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.654543 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4f4cb0dc-63c3-43d4-948a-4a355d083c2c-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-b9kd6\" (UID: \"4f4cb0dc-63c3-43d4-948a-4a355d083c2c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.679216 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/02a21759-421a-429b-ac1b-c43cd7c236f0-installation-pull-secrets\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.684078 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7ngc\" (UniqueName: \"kubernetes.io/projected/4f4cb0dc-63c3-43d4-948a-4a355d083c2c-kube-api-access-z7ngc\") pod \"package-server-manager-789f6589d5-b9kd6\" (UID: \"4f4cb0dc-63c3-43d4-948a-4a355d083c2c\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:07:37 crc kubenswrapper[4923]: W1009 10:07:37.698656 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4464ee82_fef7_46d6_8118_1e1d685e63f6.slice/crio-15bdd7615aab046d54eea25b08f54d8dfebb6056c16b798cc06a7e84d2786e4f WatchSource:0}: Error finding container 15bdd7615aab046d54eea25b08f54d8dfebb6056c16b798cc06a7e84d2786e4f: Status 404 returned error can't find the container with id 15bdd7615aab046d54eea25b08f54d8dfebb6056c16b798cc06a7e84d2786e4f Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.702628 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.703852 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5pxp\" (UniqueName: \"kubernetes.io/projected/cc177fc5-d31c-4488-ac3f-786acf59fd09-kube-api-access-t5pxp\") pod \"olm-operator-6b444d44fb-5qtv8\" (UID: \"cc177fc5-d31c-4488-ac3f-786acf59fd09\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.733803 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vh6l\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-kube-api-access-8vh6l\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755329 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755600 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-plugins-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755634 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7aba3bac-1dab-4f36-9ccf-2496c799d54d-config-volume\") pod \"dns-default-8fckr\" (UID: \"7aba3bac-1dab-4f36-9ccf-2496c799d54d\") " pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755651 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-mountpoint-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755671 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96098499-0c0f-4650-8f9f-86094e45f19a-webhook-cert\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755704 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whwqb\" (UniqueName: \"kubernetes.io/projected/7aba3bac-1dab-4f36-9ccf-2496c799d54d-kube-api-access-whwqb\") pod \"dns-default-8fckr\" (UID: \"7aba3bac-1dab-4f36-9ccf-2496c799d54d\") " pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755721 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4c6e281d-d5ff-4554-874e-a17cacea5b5a-certs\") pod \"machine-config-server-7jnjl\" (UID: \"4c6e281d-d5ff-4554-874e-a17cacea5b5a\") " pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755739 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7aba3bac-1dab-4f36-9ccf-2496c799d54d-metrics-tls\") pod \"dns-default-8fckr\" (UID: \"7aba3bac-1dab-4f36-9ccf-2496c799d54d\") " pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755772 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96098499-0c0f-4650-8f9f-86094e45f19a-apiservice-cert\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755808 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2xwm\" (UniqueName: \"kubernetes.io/projected/4c6e281d-d5ff-4554-874e-a17cacea5b5a-kube-api-access-l2xwm\") pod \"machine-config-server-7jnjl\" (UID: \"4c6e281d-d5ff-4554-874e-a17cacea5b5a\") " pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755869 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-registration-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755893 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4c6e281d-d5ff-4554-874e-a17cacea5b5a-node-bootstrap-token\") pod \"machine-config-server-7jnjl\" (UID: \"4c6e281d-d5ff-4554-874e-a17cacea5b5a\") " pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755916 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-csi-data-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755935 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwdrf\" (UniqueName: \"kubernetes.io/projected/eb5ecfa1-07aa-48ca-8727-a28a935fd937-kube-api-access-wwdrf\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.755979 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blp9l\" (UniqueName: \"kubernetes.io/projected/96098499-0c0f-4650-8f9f-86094e45f19a-kube-api-access-blp9l\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.756027 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/96098499-0c0f-4650-8f9f-86094e45f19a-tmpfs\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.756043 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-socket-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.756368 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-socket-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: E1009 10:07:37.756440 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.256423299 +0000 UTC m=+144.324605055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.756468 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-plugins-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.757093 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7aba3bac-1dab-4f36-9ccf-2496c799d54d-config-volume\") pod \"dns-default-8fckr\" (UID: \"7aba3bac-1dab-4f36-9ccf-2496c799d54d\") " pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.757140 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-mountpoint-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.760242 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-registration-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.763027 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/96098499-0c0f-4650-8f9f-86094e45f19a-tmpfs\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.763848 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4c6e281d-d5ff-4554-874e-a17cacea5b5a-node-bootstrap-token\") pod \"machine-config-server-7jnjl\" (UID: \"4c6e281d-d5ff-4554-874e-a17cacea5b5a\") " pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.764180 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/96098499-0c0f-4650-8f9f-86094e45f19a-webhook-cert\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.764467 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/eb5ecfa1-07aa-48ca-8727-a28a935fd937-csi-data-dir\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.772560 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/96098499-0c0f-4650-8f9f-86094e45f19a-apiservice-cert\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.775563 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.787043 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4c6e281d-d5ff-4554-874e-a17cacea5b5a-certs\") pod \"machine-config-server-7jnjl\" (UID: \"4c6e281d-d5ff-4554-874e-a17cacea5b5a\") " pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.789083 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7aba3bac-1dab-4f36-9ccf-2496c799d54d-metrics-tls\") pod \"dns-default-8fckr\" (UID: \"7aba3bac-1dab-4f36-9ccf-2496c799d54d\") " pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.795429 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.838826 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwdrf\" (UniqueName: \"kubernetes.io/projected/eb5ecfa1-07aa-48ca-8727-a28a935fd937-kube-api-access-wwdrf\") pod \"csi-hostpathplugin-p9k66\" (UID: \"eb5ecfa1-07aa-48ca-8727-a28a935fd937\") " pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.843378 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whwqb\" (UniqueName: \"kubernetes.io/projected/7aba3bac-1dab-4f36-9ccf-2496c799d54d-kube-api-access-whwqb\") pod \"dns-default-8fckr\" (UID: \"7aba3bac-1dab-4f36-9ccf-2496c799d54d\") " pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.848438 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blp9l\" (UniqueName: \"kubernetes.io/projected/96098499-0c0f-4650-8f9f-86094e45f19a-kube-api-access-blp9l\") pod \"packageserver-d55dfcdfc-n7fws\" (UID: \"96098499-0c0f-4650-8f9f-86094e45f19a\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.858007 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:37 crc kubenswrapper[4923]: E1009 10:07:37.860207 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.360157869 +0000 UTC m=+144.428339625 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.860520 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-p9k66" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.873230 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2xwm\" (UniqueName: \"kubernetes.io/projected/4c6e281d-d5ff-4554-874e-a17cacea5b5a-kube-api-access-l2xwm\") pod \"machine-config-server-7jnjl\" (UID: \"4c6e281d-d5ff-4554-874e-a17cacea5b5a\") " pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.938832 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.959269 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:37 crc kubenswrapper[4923]: E1009 10:07:37.959642 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.459622393 +0000 UTC m=+144.527804149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.966271 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.977308 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-5qc5h"] Oct 09 10:07:37 crc kubenswrapper[4923]: I1009 10:07:37.989695 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.062049 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.062467 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.562452699 +0000 UTC m=+144.630634455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.121028 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-7jnjl" Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.130966 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:38 crc kubenswrapper[4923]: W1009 10:07:38.134377 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3bf69a9c_5575_4c6f_8751_5ca9f3640863.slice/crio-4c4c358bc48600d7a186732ff032f6d8d06e2b133b432ef82c8c30a47d865d98 WatchSource:0}: Error finding container 4c4c358bc48600d7a186732ff032f6d8d06e2b133b432ef82c8c30a47d865d98: Status 404 returned error can't find the container with id 4c4c358bc48600d7a186732ff032f6d8d06e2b133b432ef82c8c30a47d865d98 Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.139381 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:38 crc kubenswrapper[4923]: W1009 10:07:38.144882 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44a13fe5_55be_4e50_b2c1_7d0ba939c613.slice/crio-98825a0a9353ba82e816db02c51cd0f4f32fef1454dbc2dbafcbed37627c7c69 WatchSource:0}: Error finding container 98825a0a9353ba82e816db02c51cd0f4f32fef1454dbc2dbafcbed37627c7c69: Status 404 returned error can't find the container with id 98825a0a9353ba82e816db02c51cd0f4f32fef1454dbc2dbafcbed37627c7c69 Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.164115 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.164296 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.664258636 +0000 UTC m=+144.732440392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.164570 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.165021 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.665004646 +0000 UTC m=+144.733186402 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.265888 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.266059 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.766023692 +0000 UTC m=+144.834205448 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.266665 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.267156 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.767138223 +0000 UTC m=+144.835319979 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.285911 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-5qd2g"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.367625 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.372955 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.872929778 +0000 UTC m=+144.941111534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.376453 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.377048 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.877032439 +0000 UTC m=+144.945214195 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.409805 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" event={"ID":"d755342a-48bb-467b-bc8a-5f18a8389ceb","Type":"ContainerStarted","Data":"c6df2b56fafd41449ebd593252aa45183ca84b660dba910c8ba875742a440910"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.413453 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" event={"ID":"3bf69a9c-5575-4c6f-8751-5ca9f3640863","Type":"ContainerStarted","Data":"4c4c358bc48600d7a186732ff032f6d8d06e2b133b432ef82c8c30a47d865d98"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.420205 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-44mh6"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.433390 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.436789 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.440424 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" event={"ID":"681bc3f8-88f8-4d02-ba8c-d74ee2e93058","Type":"ContainerStarted","Data":"9258c92d99e8708e735bbb6ac82b95d59e10de0e25442da2604dfd7c7f3ef0c1"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.448365 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" event={"ID":"cecd9316-6fa9-4f5f-bc1c-aa468860ea3d","Type":"ContainerStarted","Data":"9bd85922010d467f8902dd539d28c5112a1f8201c4f9ad46a867a26abcd7a87e"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.457831 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.483664 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.483994 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.983962945 +0000 UTC m=+145.052144701 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.484239 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.485993 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:38.985972479 +0000 UTC m=+145.054154425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.494375 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-2dxwn"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.569889 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.572008 4923 generic.go:334] "Generic (PLEG): container finished" podID="c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d" containerID="dbef4223d47079f27ae2a44e0e68253fa973bf07071e777105c1a36ca66ab95a" exitCode=0 Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.572131 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" event={"ID":"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d","Type":"ContainerDied","Data":"dbef4223d47079f27ae2a44e0e68253fa973bf07071e777105c1a36ca66ab95a"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.587173 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.588256 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm"] Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.588917 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:39.088870416 +0000 UTC m=+145.157052332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.674010 4923 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-d8h64 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.674096 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" podUID="45725226-1710-4204-8055-ecea69f828ee" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.693683 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.695413 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:39.195392271 +0000 UTC m=+145.263574227 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.797592 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.799668 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:39.299628724 +0000 UTC m=+145.367810540 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.817885 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qw297" podStartSLOduration=122.817859236 podStartE2EDuration="2m2.817859236s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:38.817182718 +0000 UTC m=+144.885364484" watchObservedRunningTime="2025-10-09 10:07:38.817859236 +0000 UTC m=+144.886040992" Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.826654 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.826732 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.826791 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-9rq8l"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.826824 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" event={"ID":"7eebe686-79b8-4742-bd0b-09db708418d8","Type":"ContainerStarted","Data":"374dbb45c973be7914423af6ff7d6dc44b598862626b97b3b3d39b53824fada7"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.826846 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" event={"ID":"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6","Type":"ContainerStarted","Data":"958525b905c6121da9f5c33092927fb9d9024e92437a6323393a3e5cdf6b1cb4"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.826862 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" event={"ID":"c37f83cf-f85a-4872-8e9e-ed37253c753f","Type":"ContainerStarted","Data":"2b096f91a41b5f0372c0b024f3e8772ee9010b680e12f747c524e0c1d1a45acb"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.826886 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" event={"ID":"d06b0170-f718-4285-a25b-b5c07868e3c2","Type":"ContainerStarted","Data":"8d79ae2eb217221a02322105cf6dfb8d236661b0b6f407e1f8f9348f9157ef54"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.826900 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d" event={"ID":"44a13fe5-55be-4e50-b2c1-7d0ba939c613","Type":"ContainerStarted","Data":"98825a0a9353ba82e816db02c51cd0f4f32fef1454dbc2dbafcbed37627c7c69"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.826999 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-85h6b" event={"ID":"68e9fb5c-167b-4528-9c31-97cbdc41aecd","Type":"ContainerStarted","Data":"c43c2752997f0a6df2c0965cb6eb57f2d766a103ec3909ddfa5b172745e4ad44"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.827059 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.827091 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" event={"ID":"45725226-1710-4204-8055-ecea69f828ee","Type":"ContainerStarted","Data":"9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.827103 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr"] Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.827115 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" event={"ID":"45725226-1710-4204-8055-ecea69f828ee","Type":"ContainerStarted","Data":"6b2f75cbad98fd90401800f66f8507673fc44ac6f1f00c94756370a0a2fb240b"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.827142 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" event={"ID":"4464ee82-fef7-46d6-8118-1e1d685e63f6","Type":"ContainerStarted","Data":"15bdd7615aab046d54eea25b08f54d8dfebb6056c16b798cc06a7e84d2786e4f"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.827173 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" event={"ID":"eb1393c2-0fc7-4150-a96e-d48ec1d45992","Type":"ContainerStarted","Data":"42dbe3de33ad3f759f8c3a69c0d12b0108e922bd6777cec7ee7d1825b08e7cdc"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.827184 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" event={"ID":"6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0","Type":"ContainerStarted","Data":"ee86ba9bd96b5fb3b6715f5d8421fe893456a5a2d990b63399b36eab39f5e9ce"} Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.900335 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:38 crc kubenswrapper[4923]: E1009 10:07:38.904373 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:39.40434504 +0000 UTC m=+145.472526946 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.957737 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" podStartSLOduration=122.95771725 podStartE2EDuration="2m2.95771725s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:38.957269139 +0000 UTC m=+145.025450895" watchObservedRunningTime="2025-10-09 10:07:38.95771725 +0000 UTC m=+145.025899006" Oct 09 10:07:38 crc kubenswrapper[4923]: I1009 10:07:38.987681 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-66f2z" podStartSLOduration=122.987658089 podStartE2EDuration="2m2.987658089s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:38.984445202 +0000 UTC m=+145.052626958" watchObservedRunningTime="2025-10-09 10:07:38.987658089 +0000 UTC m=+145.055839845" Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.004665 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.005067 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:39.505044318 +0000 UTC m=+145.573226074 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.108104 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.110419 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:39.610402082 +0000 UTC m=+145.678583838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.217440 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.217949 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:39.717923943 +0000 UTC m=+145.786105699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.319024 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.320071 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:39.82005279 +0000 UTC m=+145.888234546 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.364095 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.420383 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.420719 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:39.920697537 +0000 UTC m=+145.988879293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.444189 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.444286 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.512346 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" podStartSLOduration=123.512309209 podStartE2EDuration="2m3.512309209s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:39.433343788 +0000 UTC m=+145.501525554" watchObservedRunningTime="2025-10-09 10:07:39.512309209 +0000 UTC m=+145.580490965" Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.520311 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-b9mj6"] Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.538923 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.539345 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:40.039329568 +0000 UTC m=+146.107511324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.557554 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh"] Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.642230 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.643304 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:40.143276074 +0000 UTC m=+146.211457830 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.644695 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-85h6b" podStartSLOduration=123.644665952 podStartE2EDuration="2m3.644665952s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:39.593017807 +0000 UTC m=+145.661199573" watchObservedRunningTime="2025-10-09 10:07:39.644665952 +0000 UTC m=+145.712847708" Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.645587 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-897kb"] Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.706603 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4"] Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.745743 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.746355 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:40.246328095 +0000 UTC m=+146.314509921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.789052 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" event={"ID":"8abb0401-2b84-4ae7-abe1-6e151c6b1bb6","Type":"ContainerStarted","Data":"f3cd38594fc7be16717faf69a3156d49f3e8d2cf7f0ef7151421f236370d8fa9"} Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.789264 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.791416 4923 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-v2b95 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.791518 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" podUID="8abb0401-2b84-4ae7-abe1-6e151c6b1bb6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.837723 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" podStartSLOduration=123.837698691 podStartE2EDuration="2m3.837698691s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:39.83209564 +0000 UTC m=+145.900277396" watchObservedRunningTime="2025-10-09 10:07:39.837698691 +0000 UTC m=+145.905880447" Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.858387 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.859728 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:40.359712715 +0000 UTC m=+146.427894471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.877047 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc"] Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.886361 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-8n9cm"] Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.893538 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" event={"ID":"792dbe38-981d-4360-bf53-4b79947bc556","Type":"ContainerStarted","Data":"ad02c486d0abea5be5aff52eff0f161a76a8affe5c9f4193f6520e426452e8eb"} Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.898943 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" event={"ID":"401d0883-625c-4bb6-80ba-8471f93c388b","Type":"ContainerStarted","Data":"2f149ca353fa0d741e635b3c2228b379c3c5a6121d3a9aa78c4b1a3461ccaaf2"} Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.899780 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6"] Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.926704 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" event={"ID":"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e","Type":"ContainerStarted","Data":"5cdfe392cf4cec35e4a0f86788771ec9f8e5f8cecdfe71eeb35612e0047fdc73"} Oct 09 10:07:39 crc kubenswrapper[4923]: W1009 10:07:39.966005 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9302abd4_6cf6_485b_97a1_0c0805226f4c.slice/crio-ffaf05eeb61402be67a3373963d9e2b6bca47c3f9e7e33fdb7e79e31400d68ff WatchSource:0}: Error finding container ffaf05eeb61402be67a3373963d9e2b6bca47c3f9e7e33fdb7e79e31400d68ff: Status 404 returned error can't find the container with id ffaf05eeb61402be67a3373963d9e2b6bca47c3f9e7e33fdb7e79e31400d68ff Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.985272 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" event={"ID":"d06b0170-f718-4285-a25b-b5c07868e3c2","Type":"ContainerStarted","Data":"0eaf171a15c3401992f9bd78c7a36d6a445a348975768460998b0da5afde3df8"} Oct 09 10:07:39 crc kubenswrapper[4923]: I1009 10:07:39.987035 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:39 crc kubenswrapper[4923]: E1009 10:07:39.987819 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:40.487745091 +0000 UTC m=+146.555926847 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: W1009 10:07:40.015674 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcddaff14_4c07_4e69_b4be_0a16cb8db61d.slice/crio-61731d9e259df44023333a04595bea8d63b62398ec9a4dd98d72f4a231b3954a WatchSource:0}: Error finding container 61731d9e259df44023333a04595bea8d63b62398ec9a4dd98d72f4a231b3954a: Status 404 returned error can't find the container with id 61731d9e259df44023333a04595bea8d63b62398ec9a4dd98d72f4a231b3954a Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.020147 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" event={"ID":"7eebe686-79b8-4742-bd0b-09db708418d8","Type":"ContainerStarted","Data":"e7010c3b875cca4176a0117b73299d66e24bee1c09bc7056a4173733a0fe4f9e"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.039385 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9rq8l" event={"ID":"dc7bae79-3278-46f0-91fd-9a49b4057200","Type":"ContainerStarted","Data":"6f70f386a6e002fb887841f989b27ecfb94601be6406da16710cedc52dbe802b"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.042661 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8"] Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.043330 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-9sd7w" podStartSLOduration=124.04329609 podStartE2EDuration="2m4.04329609s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:40.026898588 +0000 UTC m=+146.095080344" watchObservedRunningTime="2025-10-09 10:07:40.04329609 +0000 UTC m=+146.111477856" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.066323 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5qd2g" event={"ID":"a8e9f6b9-0e46-4271-89c3-657726d9dbed","Type":"ContainerStarted","Data":"7d65214f73478a628a2e9ba0a17e2200671441ad72a82241ebc2ca96893c38d0"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.067442 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.076209 4923 patch_prober.go:28] interesting pod/console-operator-58897d9998-5qd2g container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.076262 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5qd2g" podUID="a8e9f6b9-0e46-4271-89c3-657726d9dbed" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.087006 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d" event={"ID":"44a13fe5-55be-4e50-b2c1-7d0ba939c613","Type":"ContainerStarted","Data":"0069505bd1d460703a9400f46cc0b0f72ed04a8ea4e040cdb3684842df49af50"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.089159 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.090764 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:40.59071836 +0000 UTC m=+146.658900286 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.105180 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-85h6b" event={"ID":"68e9fb5c-167b-4528-9c31-97cbdc41aecd","Type":"ContainerStarted","Data":"5670ed9a66ae7ca64f434167beb2866ecfdbb17c8b7ac3f4248a5e624de2db96"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.116883 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws"] Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.129378 4923 generic.go:334] "Generic (PLEG): container finished" podID="4464ee82-fef7-46d6-8118-1e1d685e63f6" containerID="361f7376b36b107939effa34f9286904443735033ef54a3e26e62d2b7856498c" exitCode=0 Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.131165 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2"] Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.131228 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-8fckr"] Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.131239 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" event={"ID":"4464ee82-fef7-46d6-8118-1e1d685e63f6","Type":"ContainerDied","Data":"361f7376b36b107939effa34f9286904443735033ef54a3e26e62d2b7856498c"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.131265 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-p9k66"] Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.133738 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-5qd2g" podStartSLOduration=124.133726411 podStartE2EDuration="2m4.133726411s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:40.123697421 +0000 UTC m=+146.191879177" watchObservedRunningTime="2025-10-09 10:07:40.133726411 +0000 UTC m=+146.201908167" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.178987 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7jnjl" event={"ID":"4c6e281d-d5ff-4554-874e-a17cacea5b5a","Type":"ContainerStarted","Data":"4409d52d07c35abf4e8dad05b076487ba65120f693a462390b9d889bf84fae65"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.189631 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" event={"ID":"a7c8065b-1f23-4125-9ef0-4caa50b70c4c","Type":"ContainerStarted","Data":"10bddd682230387efe95eea28689572864082735501341236a9b9d6540a7e512"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.195950 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.196584 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:40.696561607 +0000 UTC m=+146.764743363 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.197341 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" event={"ID":"a8b9202a-16ae-4569-9b82-e7342d6b82c5","Type":"ContainerStarted","Data":"9195d17c849291124720fa5485d93fa963644a1288bccef04a2d510e16421669"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.207552 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-7jnjl" podStartSLOduration=6.207527052 podStartE2EDuration="6.207527052s" podCreationTimestamp="2025-10-09 10:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:40.204227703 +0000 UTC m=+146.272409449" watchObservedRunningTime="2025-10-09 10:07:40.207527052 +0000 UTC m=+146.275708808" Oct 09 10:07:40 crc kubenswrapper[4923]: W1009 10:07:40.240713 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod96098499_0c0f_4650_8f9f_86094e45f19a.slice/crio-2d421abf80fb822c75c06f5d59b5f7c48c784daec1d9aa98578bceb52e9a769b WatchSource:0}: Error finding container 2d421abf80fb822c75c06f5d59b5f7c48c784daec1d9aa98578bceb52e9a769b: Status 404 returned error can't find the container with id 2d421abf80fb822c75c06f5d59b5f7c48c784daec1d9aa98578bceb52e9a769b Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.246380 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" event={"ID":"eb1393c2-0fc7-4150-a96e-d48ec1d45992","Type":"ContainerStarted","Data":"a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.247133 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.258164 4923 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-lwzgh container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" start-of-body= Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.258253 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" podUID="eb1393c2-0fc7-4150-a96e-d48ec1d45992" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.283892 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" podStartSLOduration=124.283876004 podStartE2EDuration="2m4.283876004s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:40.273070722 +0000 UTC m=+146.341252468" watchObservedRunningTime="2025-10-09 10:07:40.283876004 +0000 UTC m=+146.352057760" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.288885 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" event={"ID":"c37f83cf-f85a-4872-8e9e-ed37253c753f","Type":"ContainerStarted","Data":"b5ca295165c5c8bf7d7f560cfea9af8142ccd06411f6d45d9ae3ab57bc1a0b76"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.300588 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.300637 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" event={"ID":"681bc3f8-88f8-4d02-ba8c-d74ee2e93058","Type":"ContainerStarted","Data":"e2d84916c1467fc646b793938ca711aa954162f4c6296306125ff39bfb2ff0c9"} Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.302298 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:40.80227812 +0000 UTC m=+146.870459876 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.310729 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-z97pw" podStartSLOduration=124.310710077 podStartE2EDuration="2m4.310710077s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:40.309715421 +0000 UTC m=+146.377897197" watchObservedRunningTime="2025-10-09 10:07:40.310710077 +0000 UTC m=+146.378891833" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.315264 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" event={"ID":"54841e96-db7c-4118-bbfb-3d35bbf59a4d","Type":"ContainerStarted","Data":"b38cc6dd0932a03424d5bf6ee4c927ef353b1c1d0f1da2415c2a5cf297e59a6b"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.321476 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" event={"ID":"9b0d7439-9f5b-4a4f-ab08-64d77a966845","Type":"ContainerStarted","Data":"7288cc3cf7723a170e86046c67f0d7d2877081688a1c7957675c08f2e5d4c987"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.334025 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-26ghk" podStartSLOduration=125.333994936 podStartE2EDuration="2m5.333994936s" podCreationTimestamp="2025-10-09 10:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:40.333282176 +0000 UTC m=+146.401463922" watchObservedRunningTime="2025-10-09 10:07:40.333994936 +0000 UTC m=+146.402176692" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.343695 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" event={"ID":"cecd9316-6fa9-4f5f-bc1c-aa468860ea3d","Type":"ContainerStarted","Data":"b2e718bde85e8dd2cca2190b358f359a767ff9736630a26c86ec5348f13ff112"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.353544 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" event={"ID":"e32ccf22-28fb-4b8e-8e58-2789580d723c","Type":"ContainerStarted","Data":"b4efae9663a103457a1aee761715be1e3e0dd59b8c4af2fcc8342cf0aecfbb9b"} Oct 09 10:07:40 crc kubenswrapper[4923]: W1009 10:07:40.355564 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb5ecfa1_07aa_48ca_8727_a28a935fd937.slice/crio-226a1884da41a12252e5ec84a4a400dd61c207bcce067d2050d709f939ba1d88 WatchSource:0}: Error finding container 226a1884da41a12252e5ec84a4a400dd61c207bcce067d2050d709f939ba1d88: Status 404 returned error can't find the container with id 226a1884da41a12252e5ec84a4a400dd61c207bcce067d2050d709f939ba1d88 Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.379300 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" event={"ID":"d755342a-48bb-467b-bc8a-5f18a8389ceb","Type":"ContainerStarted","Data":"41d8bfa91ebcc0e503d757d812e702eed272adc6cb0bbabc90f1840d54bf5f39"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.388543 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:40 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:40 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:40 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.388607 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.396287 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" event={"ID":"815ec724-0b09-4f07-8b6e-37144aaebae0","Type":"ContainerStarted","Data":"f35eefaa16090711c46dffc10f1c9cc79d21fd06714857410cd90ffcdc85147e"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.404037 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.429614 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:40.929533205 +0000 UTC m=+146.997714961 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.431178 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" event={"ID":"981e0fe0-3fa9-403a-a147-5d31881cf83f","Type":"ContainerStarted","Data":"6e8a622c8f33a4b43889d543388e05066e2f94221db578c6025580bedc31f301"} Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.440683 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" podStartSLOduration=125.440652015 podStartE2EDuration="2m5.440652015s" podCreationTimestamp="2025-10-09 10:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:40.396693488 +0000 UTC m=+146.464875264" watchObservedRunningTime="2025-10-09 10:07:40.440652015 +0000 UTC m=+146.508833771" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.442899 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" podStartSLOduration=124.442876894 podStartE2EDuration="2m4.442876894s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:40.438140087 +0000 UTC m=+146.506321843" watchObservedRunningTime="2025-10-09 10:07:40.442876894 +0000 UTC m=+146.511058650" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.492692 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.526106 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" podStartSLOduration=124.52608746 podStartE2EDuration="2m4.52608746s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:40.488318801 +0000 UTC m=+146.556500557" watchObservedRunningTime="2025-10-09 10:07:40.52608746 +0000 UTC m=+146.594269216" Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.530649 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.532470 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.032430711 +0000 UTC m=+147.100612647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.632234 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.632680 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.132666767 +0000 UTC m=+147.200848523 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.734336 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.734536 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.234500956 +0000 UTC m=+147.302682722 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.735304 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.735655 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.235636826 +0000 UTC m=+147.303818592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.843701 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.844205 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.344184606 +0000 UTC m=+147.412366362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:40 crc kubenswrapper[4923]: I1009 10:07:40.952390 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:40 crc kubenswrapper[4923]: E1009 10:07:40.954601 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.454577945 +0000 UTC m=+147.522759701 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.059064 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.059280 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.55925411 +0000 UTC m=+147.627435866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.063313 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.064061 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.56404797 +0000 UTC m=+147.632229726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.168377 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.168901 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.668876739 +0000 UTC m=+147.737058495 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.270716 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.271106 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.771093238 +0000 UTC m=+147.839274994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.372258 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.372418 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.872393012 +0000 UTC m=+147.940574768 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.372582 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.373086 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.873047019 +0000 UTC m=+147.941228775 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.375114 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:41 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:41 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:41 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.375206 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.473860 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.475891 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:41.975861264 +0000 UTC m=+148.044043020 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.489746 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8fckr" event={"ID":"7aba3bac-1dab-4f36-9ccf-2496c799d54d","Type":"ContainerStarted","Data":"c4efa283235210778786536f8080680e27957f09d3e9c4b1c423909566a3e136"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.499677 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" event={"ID":"cc177fc5-d31c-4488-ac3f-786acf59fd09","Type":"ContainerStarted","Data":"b513752099270977565df1fdbd7306f2d5ca79500de36bf882e056e3be397586"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.499741 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" event={"ID":"cc177fc5-d31c-4488-ac3f-786acf59fd09","Type":"ContainerStarted","Data":"7ba9bce9a50d65f52060c9904c36686045f4f5dffc89aa7ab9b69d8341108b2d"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.503314 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" event={"ID":"3bf69a9c-5575-4c6f-8751-5ca9f3640863","Type":"ContainerStarted","Data":"e50c5f0da55788d21cb8d87ad84f4d02e4e54e26ad0a960d95c9dd4d005f4d42"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.582718 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" event={"ID":"7eebe686-79b8-4742-bd0b-09db708418d8","Type":"ContainerStarted","Data":"8ed01c2975e91821a6230a018a11c214796ca8e06a4091a704cfb8ed25f3102e"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.590028 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.591166 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.091146836 +0000 UTC m=+148.159328592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.620193 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" event={"ID":"e32ccf22-28fb-4b8e-8e58-2789580d723c","Type":"ContainerStarted","Data":"656b22ebc598c0a57778d41f1380cb85938d8549f444b12753cc6fb70ae6e92e"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.672227 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" event={"ID":"215543fc-99e1-47f3-9ffd-9adb4fd9c611","Type":"ContainerStarted","Data":"02cdf5e5e12375439ebc594bb8bb081f9d1e10c5e1e4d28d2eed9a3c2811d3c3"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.680383 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rvfh6" podStartSLOduration=125.680363794 podStartE2EDuration="2m5.680363794s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:41.60684923 +0000 UTC m=+147.675030996" watchObservedRunningTime="2025-10-09 10:07:41.680363794 +0000 UTC m=+147.748545550" Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.682906 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sgdx2" podStartSLOduration=125.682897802 podStartE2EDuration="2m5.682897802s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:41.682465951 +0000 UTC m=+147.750647707" watchObservedRunningTime="2025-10-09 10:07:41.682897802 +0000 UTC m=+147.751079558" Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.694526 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.697721 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.197690041 +0000 UTC m=+148.265871957 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.746555 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" event={"ID":"815ec724-0b09-4f07-8b6e-37144aaebae0","Type":"ContainerStarted","Data":"a7c3eb413d83514e0aa98a47f37e85a8945f6ea632dbc5dd6e88271672651b83"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.750110 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" podStartSLOduration=125.750096066 podStartE2EDuration="2m5.750096066s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:41.744485774 +0000 UTC m=+147.812667530" watchObservedRunningTime="2025-10-09 10:07:41.750096066 +0000 UTC m=+147.818277822" Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.770104 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" event={"ID":"d755342a-48bb-467b-bc8a-5f18a8389ceb","Type":"ContainerStarted","Data":"75831a1c4d13bd1abaff1e480725e75b2dc28f5b391e4dd7e52dee06dee4047d"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.791007 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-b9mj6" podStartSLOduration=125.790970279 podStartE2EDuration="2m5.790970279s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:41.790786734 +0000 UTC m=+147.858968500" watchObservedRunningTime="2025-10-09 10:07:41.790970279 +0000 UTC m=+147.859152035" Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.795996 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.797061 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" event={"ID":"6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0","Type":"ContainerStarted","Data":"eacfb98148733590847e302dae7b750b6a2a4d0d95d39a38d0e5c160563d936c"} Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.797565 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.297541016 +0000 UTC m=+148.365722962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.815310 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-7jnjl" event={"ID":"4c6e281d-d5ff-4554-874e-a17cacea5b5a","Type":"ContainerStarted","Data":"35b835fe56d5adf6b04ee05996cc766a755d6cec2f7c1840c47f3eba5ebc3b1a"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.825155 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" event={"ID":"96098499-0c0f-4650-8f9f-86094e45f19a","Type":"ContainerStarted","Data":"2d421abf80fb822c75c06f5d59b5f7c48c784daec1d9aa98578bceb52e9a769b"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.846411 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-5qd2g" event={"ID":"a8e9f6b9-0e46-4271-89c3-657726d9dbed","Type":"ContainerStarted","Data":"38a8422e3f6b6450798a1e132b1b757d8dd194edb1543af27cad750febaffce3"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.850237 4923 patch_prober.go:28] interesting pod/console-operator-58897d9998-5qd2g container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.863032 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-5qd2g" podUID="a8e9f6b9-0e46-4271-89c3-657726d9dbed" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.875527 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" event={"ID":"9302abd4-6cf6-485b-97a1-0c0805226f4c","Type":"ContainerStarted","Data":"ffaf05eeb61402be67a3373963d9e2b6bca47c3f9e7e33fdb7e79e31400d68ff"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.888675 4923 generic.go:334] "Generic (PLEG): container finished" podID="401d0883-625c-4bb6-80ba-8471f93c388b" containerID="a6e4b00764de98c8d4886ed8e94922efd28ab39431390f285a7350eeda332b6d" exitCode=0 Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.889087 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" event={"ID":"401d0883-625c-4bb6-80ba-8471f93c388b","Type":"ContainerDied","Data":"a6e4b00764de98c8d4886ed8e94922efd28ab39431390f285a7350eeda332b6d"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.907437 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:41 crc kubenswrapper[4923]: E1009 10:07:41.914419 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.41438219 +0000 UTC m=+148.482563956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.926159 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-h2spm" event={"ID":"981e0fe0-3fa9-403a-a147-5d31881cf83f","Type":"ContainerStarted","Data":"c1ceb7d84e6771dccce281c3cc3ccdfd4705cc70928dac1a69ad308c749a0f52"} Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.931713 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-lszpf" podStartSLOduration=125.931692717 podStartE2EDuration="2m5.931692717s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:41.924938824 +0000 UTC m=+147.993120590" watchObservedRunningTime="2025-10-09 10:07:41.931692717 +0000 UTC m=+147.999874473" Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.932351 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-gvl7r" podStartSLOduration=126.932344365 podStartE2EDuration="2m6.932344365s" podCreationTimestamp="2025-10-09 10:05:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:41.845603583 +0000 UTC m=+147.913785349" watchObservedRunningTime="2025-10-09 10:07:41.932344365 +0000 UTC m=+148.000526121" Oct 09 10:07:41 crc kubenswrapper[4923]: I1009 10:07:41.976023 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" event={"ID":"a8b9202a-16ae-4569-9b82-e7342d6b82c5","Type":"ContainerStarted","Data":"7e14c1623b8646399219c3f528d758c3fe06a0733ee06bde5e2c9d63a65c5790"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.013209 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.022205 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.522167089 +0000 UTC m=+148.590349045 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.052779 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" event={"ID":"9d11e36b-e397-47e6-a3d1-93cf9367f83f","Type":"ContainerStarted","Data":"7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.053321 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.053341 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" event={"ID":"9d11e36b-e397-47e6-a3d1-93cf9367f83f","Type":"ContainerStarted","Data":"39816ad0baa2267a1a0eb13eb5ce7a75744746c39da2fb86aaccda1057b610b7"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.054938 4923 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-897kb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.055002 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" podUID="9d11e36b-e397-47e6-a3d1-93cf9367f83f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.115362 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.115489 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.615465688 +0000 UTC m=+148.683647444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.115790 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.117385 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.617377308 +0000 UTC m=+148.685559064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.137153 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" event={"ID":"4f4cb0dc-63c3-43d4-948a-4a355d083c2c","Type":"ContainerStarted","Data":"030a0a521201550fe7250fa8f9826058f8033cf1aff924f4f6e3454f12f72c83"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.137235 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" event={"ID":"4f4cb0dc-63c3-43d4-948a-4a355d083c2c","Type":"ContainerStarted","Data":"fc77c6e30eafaaf62ff81ecc5ab83bcfd9eb6b317feb1c49044acc7a0ce244f1"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.138149 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.169126 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-44mh6" podStartSLOduration=126.169094055 podStartE2EDuration="2m6.169094055s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.090067281 +0000 UTC m=+148.158249037" watchObservedRunningTime="2025-10-09 10:07:42.169094055 +0000 UTC m=+148.237275811" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.170233 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-9rq8l" event={"ID":"dc7bae79-3278-46f0-91fd-9a49b4057200","Type":"ContainerStarted","Data":"4324e0a3472fe635153f5af046de577a2e04cdb45b51b0159cef5f6130ea3e6e"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.170586 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" podStartSLOduration=126.170578654 podStartE2EDuration="2m6.170578654s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.162915587 +0000 UTC m=+148.231097343" watchObservedRunningTime="2025-10-09 10:07:42.170578654 +0000 UTC m=+148.238760400" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.172020 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-9rq8l" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.191961 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-9rq8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.192022 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9rq8l" podUID="dc7bae79-3278-46f0-91fd-9a49b4057200" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.203501 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8n9cm" event={"ID":"cddaff14-4c07-4e69-b4be-0a16cb8db61d","Type":"ContainerStarted","Data":"61731d9e259df44023333a04595bea8d63b62398ec9a4dd98d72f4a231b3954a"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.217398 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.218443 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.718421606 +0000 UTC m=+148.786603362 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.249886 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" event={"ID":"c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d","Type":"ContainerStarted","Data":"36b14c73d50a5284ead1f4f1304534cf4c29cabda747e3d38acbd3a25e852366"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.250808 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.267007 4923 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-vmrkj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.267115 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" podUID="c58bcb2b-4c0a-42f7-9260-f45f96a3dd3d" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.268364 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" podStartSLOduration=126.268337713 podStartE2EDuration="2m6.268337713s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.23450614 +0000 UTC m=+148.302687896" watchObservedRunningTime="2025-10-09 10:07:42.268337713 +0000 UTC m=+148.336519469" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.289849 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" event={"ID":"a7c8065b-1f23-4125-9ef0-4caa50b70c4c","Type":"ContainerStarted","Data":"06b6d2c47f39101ad417c841d2eddbff83615d088ca7c9ff68833dac92fcb366"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.295997 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-8n9cm" podStartSLOduration=8.295976969 podStartE2EDuration="8.295976969s" podCreationTimestamp="2025-10-09 10:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.268080816 +0000 UTC m=+148.336262572" watchObservedRunningTime="2025-10-09 10:07:42.295976969 +0000 UTC m=+148.364158725" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.324200 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" podStartSLOduration=126.3241592 podStartE2EDuration="2m6.3241592s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.320911152 +0000 UTC m=+148.389092908" watchObservedRunningTime="2025-10-09 10:07:42.3241592 +0000 UTC m=+148.392340956" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.325067 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-zrf6j" event={"ID":"9b0d7439-9f5b-4a4f-ab08-64d77a966845","Type":"ContainerStarted","Data":"022a8764f3d716f13548ca3fd814592c68605fd991c039bb212e32cd6e5b1e4b"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.326938 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.328905 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.828884997 +0000 UTC m=+148.897066743 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.333494 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-9rq8l" podStartSLOduration=126.33343758 podStartE2EDuration="2m6.33343758s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.296475822 +0000 UTC m=+148.364657598" watchObservedRunningTime="2025-10-09 10:07:42.33343758 +0000 UTC m=+148.401619336" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.367168 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d" event={"ID":"44a13fe5-55be-4e50-b2c1-7d0ba939c613","Type":"ContainerStarted","Data":"11dcbbc90df357ed7da4c37a7c718b7baa5e76ecc47e7631c23e864524ac3f42"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.375056 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:42 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:42 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:42 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.375134 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.404801 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-2456d" podStartSLOduration=126.404782766 podStartE2EDuration="2m6.404782766s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.404233691 +0000 UTC m=+148.472415447" watchObservedRunningTime="2025-10-09 10:07:42.404782766 +0000 UTC m=+148.472964522" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.406870 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-jrmhr" podStartSLOduration=126.406863112 podStartE2EDuration="2m6.406863112s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.362382942 +0000 UTC m=+148.430564708" watchObservedRunningTime="2025-10-09 10:07:42.406863112 +0000 UTC m=+148.475044868" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.416213 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" event={"ID":"54841e96-db7c-4118-bbfb-3d35bbf59a4d","Type":"ContainerStarted","Data":"a5cddcb6ee7ef460b3763942f86524d8e730b1204733eeaec31044af806f1adb"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.430716 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" event={"ID":"9bc36fc2-f4d0-4381-b766-43da3213c356","Type":"ContainerStarted","Data":"95c2c21ab2a82c56e10631187a3dcec3b0d582950ad0bac7e2bc0cd82bb36e86"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.430788 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" event={"ID":"9bc36fc2-f4d0-4381-b766-43da3213c356","Type":"ContainerStarted","Data":"ea9fcc4b87a7d0f731563c1f17fdec8a80cbff95a8c65d9629cea71f9051f7a5"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.431869 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.433436 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:42.933410908 +0000 UTC m=+149.001592664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.468527 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p9k66" event={"ID":"eb5ecfa1-07aa-48ca-8727-a28a935fd937","Type":"ContainerStarted","Data":"226a1884da41a12252e5ec84a4a400dd61c207bcce067d2050d709f939ba1d88"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.502079 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-wb9v2" event={"ID":"cecd9316-6fa9-4f5f-bc1c-aa468860ea3d","Type":"ContainerStarted","Data":"a6db071f0667fe0482598f23e977c287915946a50cb01cf507b9984309f374dc"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.502462 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" podStartSLOduration=126.502432081 podStartE2EDuration="2m6.502432081s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.46570163 +0000 UTC m=+148.533883386" watchObservedRunningTime="2025-10-09 10:07:42.502432081 +0000 UTC m=+148.570613837" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.502572 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2zlvc" podStartSLOduration=126.502568135 podStartE2EDuration="2m6.502568135s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.498431653 +0000 UTC m=+148.566613409" watchObservedRunningTime="2025-10-09 10:07:42.502568135 +0000 UTC m=+148.570749901" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.528406 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" event={"ID":"792dbe38-981d-4360-bf53-4b79947bc556","Type":"ContainerStarted","Data":"63be30152178fb05f56ce2c49612fcf7417b339f00324942c81e6fb6fe5e72b9"} Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.534995 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.535478 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.035456932 +0000 UTC m=+149.103638688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.549683 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.556922 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v2b95" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.581342 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" podStartSLOduration=126.58131683 podStartE2EDuration="2m6.58131683s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:42.571052613 +0000 UTC m=+148.639234369" watchObservedRunningTime="2025-10-09 10:07:42.58131683 +0000 UTC m=+148.649498586" Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.642373 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.644088 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.144068514 +0000 UTC m=+149.212250270 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.744787 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.745259 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.245238384 +0000 UTC m=+149.313420330 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.846860 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.847451 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.347433303 +0000 UTC m=+149.415615059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:42 crc kubenswrapper[4923]: I1009 10:07:42.949545 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:42 crc kubenswrapper[4923]: E1009 10:07:42.950085 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.450062362 +0000 UTC m=+149.518244318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.051183 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.051446 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.551414018 +0000 UTC m=+149.619595774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.052034 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.052608 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.552585209 +0000 UTC m=+149.620766965 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.152862 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.153553 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.653524444 +0000 UTC m=+149.721706200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.254920 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.255404 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.755387213 +0000 UTC m=+149.823568969 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.356738 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.357225 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.857201671 +0000 UTC m=+149.925383427 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.373659 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:43 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:43 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:43 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.373794 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.459256 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.459339 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.459374 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.459452 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.460424 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:43.960403617 +0000 UTC m=+150.028585373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.463550 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.467684 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.468655 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.531640 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-vmrkj" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.536967 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8fckr" event={"ID":"7aba3bac-1dab-4f36-9ccf-2496c799d54d","Type":"ContainerStarted","Data":"08d76b16f575122bed96bcd31e9147426c32f2531152f594751113a998e3d76c"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.537038 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-8fckr" event={"ID":"7aba3bac-1dab-4f36-9ccf-2496c799d54d","Type":"ContainerStarted","Data":"d7ff70203f84ead76b7adb038517066131164669e0795b5b04b97ebdcb4449d7"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.537451 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.539910 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" event={"ID":"4f4cb0dc-63c3-43d4-948a-4a355d083c2c","Type":"ContainerStarted","Data":"78625e32d42aa578fb9de2d870f2e2131242c69b9b87042cb4a4b8bee041c842"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.544030 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" event={"ID":"4464ee82-fef7-46d6-8118-1e1d685e63f6","Type":"ContainerStarted","Data":"d1360ce8b99aa0860ca37d931f0e319e13a9fcfb4fb1f4f47b526f048ce2cea8"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.544081 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" event={"ID":"4464ee82-fef7-46d6-8118-1e1d685e63f6","Type":"ContainerStarted","Data":"b2c5ab41aeb84686cdd25653c0a07fa3bc81a1a753df3d4b49eb2b0112598742"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.545391 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" event={"ID":"215543fc-99e1-47f3-9ffd-9adb4fd9c611","Type":"ContainerStarted","Data":"d494dc0a2d9cf9f20441b740e4d1a3fb51473fb0a9ed890bfb6b0da48032c79b"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.547597 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" event={"ID":"9302abd4-6cf6-485b-97a1-0c0805226f4c","Type":"ContainerStarted","Data":"8c67697a32379c926dfea61a7cd763ad3b989291c03f78c018421d97aad70f36"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.547660 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" event={"ID":"9302abd4-6cf6-485b-97a1-0c0805226f4c","Type":"ContainerStarted","Data":"95c562981711c95ab4023f5d35ab58689f15895deae9099c67c1745f21a0b234"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.550526 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-2dxwn" event={"ID":"54841e96-db7c-4118-bbfb-3d35bbf59a4d","Type":"ContainerStarted","Data":"8148e5914e0262d160de261ff20c8798f047ddc80f3ff59946e9c60256a659d8"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.556973 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" event={"ID":"0f0da29a-a520-4aa8-ae7a-47a22fc9a81e","Type":"ContainerStarted","Data":"d2d206d0657b411b6c45d49436309c69efebb4a10cc0c2a12adca749320c1ebd"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.562671 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.562990 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.563153 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:44.063118738 +0000 UTC m=+150.131300494 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.570114 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.582937 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" event={"ID":"401d0883-625c-4bb6-80ba-8471f93c388b","Type":"ContainerStarted","Data":"44e766109dee79916c2af81ebd456f5710627a7c3e98385e236287fe204e353b"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.617957 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.620101 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" event={"ID":"96098499-0c0f-4650-8f9f-86094e45f19a","Type":"ContainerStarted","Data":"2afb3598a357c02cafa64d2dc9a1f244e2f485eff762a1786f17387fc74ad093"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.620822 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.626659 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.637789 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-d4vgh" event={"ID":"792dbe38-981d-4360-bf53-4b79947bc556","Type":"ContainerStarted","Data":"206afc5cf7bb84e424880e2b373d2cab0109e05774360b4d2c28a5b8511d3fc5"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.648143 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.652898 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p9k66" event={"ID":"eb5ecfa1-07aa-48ca-8727-a28a935fd937","Type":"ContainerStarted","Data":"ff5ec6f65490d95d31e45d91cbf48a81b1d5d6a94af98cbc8436f63d830d8219"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.659231 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" event={"ID":"6ad90a4b-f050-4f5e-a6d7-a6eeed4db2f0","Type":"ContainerStarted","Data":"76e7396a102707cd9eb6b25702c8beac763a36c9e72f6a7635393036cc5a093d"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.664494 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-8n9cm" event={"ID":"cddaff14-4c07-4e69-b4be-0a16cb8db61d","Type":"ContainerStarted","Data":"af1421943e5ffeb71bafa51e2bcebc599815ad21a268bb9b1307d921baf17f26"} Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.664552 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.666386 4923 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-897kb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" start-of-body= Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.666464 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" podUID="9d11e36b-e397-47e6-a3d1-93cf9367f83f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.27:8080/healthz\": dial tcp 10.217.0.27:8080: connect: connection refused" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.666562 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.668966 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:44.168947955 +0000 UTC m=+150.237129711 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.671214 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-9rq8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.671305 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9rq8l" podUID="dc7bae79-3278-46f0-91fd-9a49b4057200" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.684581 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-5qd2g" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.718202 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.768290 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.769363 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:44.269345065 +0000 UTC m=+150.337526901 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.882618 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.883286 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:44.38326225 +0000 UTC m=+150.451444006 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.913477 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" podStartSLOduration=127.913447384 podStartE2EDuration="2m7.913447384s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:43.899094966 +0000 UTC m=+149.967276712" watchObservedRunningTime="2025-10-09 10:07:43.913447384 +0000 UTC m=+149.981629140" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.916033 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-8fckr" podStartSLOduration=9.916018463 podStartE2EDuration="9.916018463s" podCreationTimestamp="2025-10-09 10:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:43.718094382 +0000 UTC m=+149.786276138" watchObservedRunningTime="2025-10-09 10:07:43.916018463 +0000 UTC m=+149.984200219" Oct 09 10:07:43 crc kubenswrapper[4923]: I1009 10:07:43.985851 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:43 crc kubenswrapper[4923]: E1009 10:07:43.986291 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:44.486263219 +0000 UTC m=+150.554444975 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.055591 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-ndvn4" podStartSLOduration=128.055555999 podStartE2EDuration="2m8.055555999s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:44.040918994 +0000 UTC m=+150.109100740" watchObservedRunningTime="2025-10-09 10:07:44.055555999 +0000 UTC m=+150.123737755" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.100097 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r7pbq"] Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.102600 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.105257 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:44 crc kubenswrapper[4923]: E1009 10:07:44.105735 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:44.605718413 +0000 UTC m=+150.673900169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.109870 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.242405 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.242815 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-catalog-content\") pod \"certified-operators-r7pbq\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.242958 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-utilities\") pod \"certified-operators-r7pbq\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.242993 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmhct\" (UniqueName: \"kubernetes.io/projected/39ceb27f-3ad7-452a-bbdd-92057099d011-kube-api-access-qmhct\") pod \"certified-operators-r7pbq\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: E1009 10:07:44.243207 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:44.743187323 +0000 UTC m=+150.811369079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.244563 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r7pbq"] Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.264880 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b8fd8"] Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.266387 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.267313 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6jfbh" podStartSLOduration=128.267280944 podStartE2EDuration="2m8.267280944s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:44.244486588 +0000 UTC m=+150.312668344" watchObservedRunningTime="2025-10-09 10:07:44.267280944 +0000 UTC m=+150.335462700" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.280682 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.293853 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b8fd8"] Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.311436 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jbd8g"] Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.312715 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.319806 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-5qtv8" podStartSLOduration=128.31977277 podStartE2EDuration="2m8.31977277s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:44.281063176 +0000 UTC m=+150.349244932" watchObservedRunningTime="2025-10-09 10:07:44.31977277 +0000 UTC m=+150.387954526" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.339331 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jbd8g"] Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.352466 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl8tr\" (UniqueName: \"kubernetes.io/projected/d4a3f04f-d392-4065-896b-7db7683f50e6-kube-api-access-kl8tr\") pod \"community-operators-b8fd8\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.352518 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-catalog-content\") pod \"community-operators-b8fd8\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.352544 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.352573 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-utilities\") pod \"community-operators-b8fd8\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.352595 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-utilities\") pod \"certified-operators-r7pbq\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.352614 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmhct\" (UniqueName: \"kubernetes.io/projected/39ceb27f-3ad7-452a-bbdd-92057099d011-kube-api-access-qmhct\") pod \"certified-operators-r7pbq\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.352677 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-catalog-content\") pod \"certified-operators-r7pbq\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.353512 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-catalog-content\") pod \"certified-operators-r7pbq\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: E1009 10:07:44.353883 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:44.853866341 +0000 UTC m=+150.922048097 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.354870 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" podStartSLOduration=128.354833597 podStartE2EDuration="2m8.354833597s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:44.331766324 +0000 UTC m=+150.399948080" watchObservedRunningTime="2025-10-09 10:07:44.354833597 +0000 UTC m=+150.423015353" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.355284 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-utilities\") pod \"certified-operators-r7pbq\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.370212 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:44 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:44 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:44 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.370632 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.382881 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" podStartSLOduration=128.382861423 podStartE2EDuration="2m8.382861423s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:44.380601222 +0000 UTC m=+150.448782968" watchObservedRunningTime="2025-10-09 10:07:44.382861423 +0000 UTC m=+150.451043179" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.424413 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmhct\" (UniqueName: \"kubernetes.io/projected/39ceb27f-3ad7-452a-bbdd-92057099d011-kube-api-access-qmhct\") pod \"certified-operators-r7pbq\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.460797 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.461042 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkwhc\" (UniqueName: \"kubernetes.io/projected/31d5c282-dd47-420c-bc6e-cfcbee938bfc-kube-api-access-xkwhc\") pod \"certified-operators-jbd8g\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.461079 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl8tr\" (UniqueName: \"kubernetes.io/projected/d4a3f04f-d392-4065-896b-7db7683f50e6-kube-api-access-kl8tr\") pod \"community-operators-b8fd8\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.461100 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-catalog-content\") pod \"community-operators-b8fd8\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.461130 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-utilities\") pod \"community-operators-b8fd8\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.461173 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-catalog-content\") pod \"certified-operators-jbd8g\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.461212 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-utilities\") pod \"certified-operators-jbd8g\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: E1009 10:07:44.461639 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:44.961615519 +0000 UTC m=+151.029797275 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.462452 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-catalog-content\") pod \"community-operators-b8fd8\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.462703 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-utilities\") pod \"community-operators-b8fd8\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.483531 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-n7fws" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.485183 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-qhv8h"] Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.486301 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.496283 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-5qc5h" podStartSLOduration=128.496256103 podStartE2EDuration="2m8.496256103s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:44.495555484 +0000 UTC m=+150.563737240" watchObservedRunningTime="2025-10-09 10:07:44.496256103 +0000 UTC m=+150.564437859" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.505169 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl8tr\" (UniqueName: \"kubernetes.io/projected/d4a3f04f-d392-4065-896b-7db7683f50e6-kube-api-access-kl8tr\") pod \"community-operators-b8fd8\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.526236 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qhv8h"] Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.544492 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.562784 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkwhc\" (UniqueName: \"kubernetes.io/projected/31d5c282-dd47-420c-bc6e-cfcbee938bfc-kube-api-access-xkwhc\") pod \"certified-operators-jbd8g\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.563071 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.563157 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-utilities\") pod \"community-operators-qhv8h\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.563243 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-catalog-content\") pod \"community-operators-qhv8h\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.563338 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-catalog-content\") pod \"certified-operators-jbd8g\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.563421 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-utilities\") pod \"certified-operators-jbd8g\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.563497 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp2jb\" (UniqueName: \"kubernetes.io/projected/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-kube-api-access-mp2jb\") pod \"community-operators-qhv8h\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: E1009 10:07:44.564194 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.064180697 +0000 UTC m=+151.132362453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.564839 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-catalog-content\") pod \"certified-operators-jbd8g\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.565131 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-utilities\") pod \"certified-operators-jbd8g\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.607802 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkwhc\" (UniqueName: \"kubernetes.io/projected/31d5c282-dd47-420c-bc6e-cfcbee938bfc-kube-api-access-xkwhc\") pod \"certified-operators-jbd8g\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.632088 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.666603 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.666927 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-utilities\") pod \"community-operators-qhv8h\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.666976 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-catalog-content\") pod \"community-operators-qhv8h\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.667065 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp2jb\" (UniqueName: \"kubernetes.io/projected/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-kube-api-access-mp2jb\") pod \"community-operators-qhv8h\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: E1009 10:07:44.667607 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.167581428 +0000 UTC m=+151.235763184 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.668104 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-utilities\") pod \"community-operators-qhv8h\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.668394 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-catalog-content\") pod \"community-operators-qhv8h\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.672353 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.739014 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-9rq8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.739107 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9rq8l" podUID="dc7bae79-3278-46f0-91fd-9a49b4057200" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.740343 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp2jb\" (UniqueName: \"kubernetes.io/projected/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-kube-api-access-mp2jb\") pod \"community-operators-qhv8h\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.790528 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:44 crc kubenswrapper[4923]: E1009 10:07:44.791685 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.291668267 +0000 UTC m=+151.359850023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.863377 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:07:44 crc kubenswrapper[4923]: W1009 10:07:44.873006 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-8fb77d3dfff4cbac65822b19b625068dd4af1734bb2239a82d0f50314fb55724 WatchSource:0}: Error finding container 8fb77d3dfff4cbac65822b19b625068dd4af1734bb2239a82d0f50314fb55724: Status 404 returned error can't find the container with id 8fb77d3dfff4cbac65822b19b625068dd4af1734bb2239a82d0f50314fb55724 Oct 09 10:07:44 crc kubenswrapper[4923]: I1009 10:07:44.895971 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:44 crc kubenswrapper[4923]: E1009 10:07:44.923705 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.423654149 +0000 UTC m=+151.491835915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.000140 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.000546 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.500532623 +0000 UTC m=+151.568714389 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.101311 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.101609 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.60158017 +0000 UTC m=+151.669761926 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.102924 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.103251 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.603240496 +0000 UTC m=+151.671422242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.203896 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.204388 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.704370685 +0000 UTC m=+151.772552431 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.309631 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.310389 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.810374216 +0000 UTC m=+151.878555972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.387819 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:45 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:45 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:45 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.387858 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.398540 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b8fd8"] Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.412910 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.413370 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:45.913351536 +0000 UTC m=+151.981533292 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.430929 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r7pbq"] Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.508197 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jbd8g"] Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.517779 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.520544 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.020024755 +0000 UTC m=+152.088206511 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.619310 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.620290 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.120271261 +0000 UTC m=+152.188453017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.638505 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-qhv8h"] Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.721975 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.722352 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.222334915 +0000 UTC m=+152.290516671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.755121 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p9k66" event={"ID":"eb5ecfa1-07aa-48ca-8727-a28a935fd937","Type":"ContainerStarted","Data":"91b1ae0fb8ca7c1859ea6cd19c2cf4c6f08bdfd60d8551d8a56a2b13d10a6b6e"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.756109 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbd8g" event={"ID":"31d5c282-dd47-420c-bc6e-cfcbee938bfc","Type":"ContainerStarted","Data":"ecaafaa005b7769059cf20724c96c913706cc23a9fddeeb258922d098dce9f30"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.756818 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhv8h" event={"ID":"e46ac356-170c-4c91-b8b6-36ec0dfb93d8","Type":"ContainerStarted","Data":"0ad6968e5da93c1944dbe9182b3f477fb5e125ecb148ece9fae8db40d050137f"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.759343 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7pbq" event={"ID":"39ceb27f-3ad7-452a-bbdd-92057099d011","Type":"ContainerStarted","Data":"859f2bd01c90c5a8a4c36e115e64b6e118b66065bd007cc9ce18f70bbe76218e"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.768382 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"350fea931eb896b8795597e49612102653bc3347bf262a08619552f6e76b013c"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.768448 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8fb77d3dfff4cbac65822b19b625068dd4af1734bb2239a82d0f50314fb55724"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.791890 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8fd8" event={"ID":"d4a3f04f-d392-4065-896b-7db7683f50e6","Type":"ContainerStarted","Data":"64dc7a263529a16a4fed7c0875b966a524a42a262bc8308b578982fab80df067"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.802949 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3c775b51ac919c765b60f9bd65916e2cbeb71c773f291737fc7769aaab03b16a"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.803021 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"3c506e52cfec11c2a5bef96b5ce3fd82d9ba624f759de6adba3175c694538029"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.804028 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.818402 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"6b69187143f33f78dd419545d967f0f6c65e6d5b233be4ad10c7da8155d950eb"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.818474 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"994f809278bc6a5e892272af1f3c541ce9a322b9a792f161671fc050528abd0b"} Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.838185 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.842101 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.342018465 +0000 UTC m=+152.410200221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.856876 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vs8dt"] Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.860707 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.868112 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.874566 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vs8dt"] Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.940704 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.940868 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-utilities\") pod \"redhat-marketplace-vs8dt\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.940892 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f25d5\" (UniqueName: \"kubernetes.io/projected/09e8da7b-adaf-49f9-9811-9ca90e2bff56-kube-api-access-f25d5\") pod \"redhat-marketplace-vs8dt\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:45 crc kubenswrapper[4923]: I1009 10:07:45.940950 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-catalog-content\") pod \"redhat-marketplace-vs8dt\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:45 crc kubenswrapper[4923]: E1009 10:07:45.941495 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.441467169 +0000 UTC m=+152.509648925 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.042115 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.042299 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.542263159 +0000 UTC m=+152.610444915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.042515 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.042581 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-utilities\") pod \"redhat-marketplace-vs8dt\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.042611 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f25d5\" (UniqueName: \"kubernetes.io/projected/09e8da7b-adaf-49f9-9811-9ca90e2bff56-kube-api-access-f25d5\") pod \"redhat-marketplace-vs8dt\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.042673 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-catalog-content\") pod \"redhat-marketplace-vs8dt\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.043198 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-utilities\") pod \"redhat-marketplace-vs8dt\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.042876 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.542869615 +0000 UTC m=+152.611051371 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.043400 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-catalog-content\") pod \"redhat-marketplace-vs8dt\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.063613 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f25d5\" (UniqueName: \"kubernetes.io/projected/09e8da7b-adaf-49f9-9811-9ca90e2bff56-kube-api-access-f25d5\") pod \"redhat-marketplace-vs8dt\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.115563 4923 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.144406 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.144693 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.644652963 +0000 UTC m=+152.712834729 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.144836 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.145336 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.645325111 +0000 UTC m=+152.713506867 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.188269 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.246249 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.247216 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.74719003 +0000 UTC m=+152.815371786 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.247464 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zkxzd"] Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.249606 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.266644 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkxzd"] Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.348858 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-utilities\") pod \"redhat-marketplace-zkxzd\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.348924 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.348953 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-catalog-content\") pod \"redhat-marketplace-zkxzd\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.349207 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4pj8\" (UniqueName: \"kubernetes.io/projected/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-kube-api-access-p4pj8\") pod \"redhat-marketplace-zkxzd\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.349482 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.849462571 +0000 UTC m=+152.917644327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.369028 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:46 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:46 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:46 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.369123 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.465543 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.465818 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-catalog-content\") pod \"redhat-marketplace-zkxzd\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.465895 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4pj8\" (UniqueName: \"kubernetes.io/projected/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-kube-api-access-p4pj8\") pod \"redhat-marketplace-zkxzd\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.465966 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-utilities\") pod \"redhat-marketplace-zkxzd\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.470451 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:46.970430275 +0000 UTC m=+153.038612031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.477140 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-utilities\") pod \"redhat-marketplace-zkxzd\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.479065 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-catalog-content\") pod \"redhat-marketplace-zkxzd\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.489375 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.489443 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.495158 4923 patch_prober.go:28] interesting pod/console-f9d7485db-66f2z container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.495225 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-66f2z" podUID="a8c487fa-3436-4cc9-83b3-de49256e97c9" containerName="console" probeResult="failure" output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.499260 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4pj8\" (UniqueName: \"kubernetes.io/projected/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-kube-api-access-p4pj8\") pod \"redhat-marketplace-zkxzd\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.543474 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vs8dt"] Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.568305 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.568901 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:47.068877202 +0000 UTC m=+153.137058958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.612600 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:07:46 crc kubenswrapper[4923]: W1009 10:07:46.613142 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09e8da7b_adaf_49f9_9811_9ca90e2bff56.slice/crio-277cc13a36b5fcf09a91e275a89fb9fb167e1206b4e2d28f8354b0611c351af2 WatchSource:0}: Error finding container 277cc13a36b5fcf09a91e275a89fb9fb167e1206b4e2d28f8354b0611c351af2: Status 404 returned error can't find the container with id 277cc13a36b5fcf09a91e275a89fb9fb167e1206b4e2d28f8354b0611c351af2 Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.669360 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.669652 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:47.169622261 +0000 UTC m=+153.237804027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.670112 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.670918 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:47.170902977 +0000 UTC m=+153.239084723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.772084 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.772268 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 10:07:47.272219112 +0000 UTC m=+153.340400868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.773883 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:46 crc kubenswrapper[4923]: E1009 10:07:46.774249 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 10:07:47.274233157 +0000 UTC m=+153.342414903 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-pvsh5" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.780814 4923 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-09T10:07:46.115606569Z","Handler":null,"Name":""} Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.791184 4923 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.791246 4923 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.840169 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vs8dt" event={"ID":"09e8da7b-adaf-49f9-9811-9ca90e2bff56","Type":"ContainerStarted","Data":"277cc13a36b5fcf09a91e275a89fb9fb167e1206b4e2d28f8354b0611c351af2"} Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.843176 4923 generic.go:334] "Generic (PLEG): container finished" podID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerID="ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e" exitCode=0 Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.843512 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7pbq" event={"ID":"39ceb27f-3ad7-452a-bbdd-92057099d011","Type":"ContainerDied","Data":"ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e"} Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.845444 4923 generic.go:334] "Generic (PLEG): container finished" podID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerID="8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2" exitCode=0 Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.845557 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8fd8" event={"ID":"d4a3f04f-d392-4065-896b-7db7683f50e6","Type":"ContainerDied","Data":"8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2"} Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.846342 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.850226 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p9k66" event={"ID":"eb5ecfa1-07aa-48ca-8727-a28a935fd937","Type":"ContainerStarted","Data":"5252fff90e144caeecf982015853b8ecdce252d4070701dcbeab59dae24c591d"} Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.855769 4923 generic.go:334] "Generic (PLEG): container finished" podID="215543fc-99e1-47f3-9ffd-9adb4fd9c611" containerID="d494dc0a2d9cf9f20441b740e4d1a3fb51473fb0a9ed890bfb6b0da48032c79b" exitCode=0 Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.855889 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" event={"ID":"215543fc-99e1-47f3-9ffd-9adb4fd9c611","Type":"ContainerDied","Data":"d494dc0a2d9cf9f20441b740e4d1a3fb51473fb0a9ed890bfb6b0da48032c79b"} Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.859854 4923 generic.go:334] "Generic (PLEG): container finished" podID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerID="7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05" exitCode=0 Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.859970 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhv8h" event={"ID":"e46ac356-170c-4c91-b8b6-36ec0dfb93d8","Type":"ContainerDied","Data":"7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05"} Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.876434 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.878193 4923 generic.go:334] "Generic (PLEG): container finished" podID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerID="001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d" exitCode=0 Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.878862 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbd8g" event={"ID":"31d5c282-dd47-420c-bc6e-cfcbee938bfc","Type":"ContainerDied","Data":"001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d"} Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.881833 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.946297 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.947045 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.964718 4923 patch_prober.go:28] interesting pod/apiserver-76f77b778f-dtw6g container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]log ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]etcd ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/start-apiserver-admission-initializer ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/generic-apiserver-start-informers ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/max-in-flight-filter ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/storage-object-count-tracker-hook ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/image.openshift.io-apiserver-caches ok Oct 09 10:07:46 crc kubenswrapper[4923]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Oct 09 10:07:46 crc kubenswrapper[4923]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/project.openshift.io-projectcache ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/openshift.io-startinformers ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/openshift.io-restmapperupdater ok Oct 09 10:07:46 crc kubenswrapper[4923]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Oct 09 10:07:46 crc kubenswrapper[4923]: livez check failed Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.964832 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" podUID="4464ee82-fef7-46d6-8118-1e1d685e63f6" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.979735 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.993462 4923 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 09 10:07:46 crc kubenswrapper[4923]: I1009 10:07:46.993509 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.041563 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-pvsh5\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.062461 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nmsxl"] Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.064441 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.068142 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.070830 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nmsxl"] Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.137514 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkxzd"] Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.167327 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.167381 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.177670 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.185579 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-catalog-content\") pod \"redhat-operators-nmsxl\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.185649 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xksvv\" (UniqueName: \"kubernetes.io/projected/fa810245-ceec-4d06-b02d-20e39d65ff1f-kube-api-access-xksvv\") pod \"redhat-operators-nmsxl\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.185682 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-utilities\") pod \"redhat-operators-nmsxl\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.286882 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-catalog-content\") pod \"redhat-operators-nmsxl\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.286969 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xksvv\" (UniqueName: \"kubernetes.io/projected/fa810245-ceec-4d06-b02d-20e39d65ff1f-kube-api-access-xksvv\") pod \"redhat-operators-nmsxl\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.286997 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-utilities\") pod \"redhat-operators-nmsxl\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.287902 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-utilities\") pod \"redhat-operators-nmsxl\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.288056 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-catalog-content\") pod \"redhat-operators-nmsxl\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.318886 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xksvv\" (UniqueName: \"kubernetes.io/projected/fa810245-ceec-4d06-b02d-20e39d65ff1f-kube-api-access-xksvv\") pod \"redhat-operators-nmsxl\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.333678 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.361611 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.370921 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.373816 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:47 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:47 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:47 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.373895 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.375273 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.383347 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.383983 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.388448 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.412077 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.436376 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-9rq8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.436454 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-9rq8l container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.436467 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9rq8l" podUID="dc7bae79-3278-46f0-91fd-9a49b4057200" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.436528 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9rq8l" podUID="dc7bae79-3278-46f0-91fd-9a49b4057200" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.440218 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-stwns"] Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.442244 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.467954 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-stwns"] Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.469222 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.489270 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.489412 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.591220 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.591933 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.591867 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22dv4\" (UniqueName: \"kubernetes.io/projected/66135764-cd30-4365-bc33-c1fb96eedc85-kube-api-access-22dv4\") pod \"redhat-operators-stwns\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.592141 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.592215 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-catalog-content\") pod \"redhat-operators-stwns\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.592282 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-utilities\") pod \"redhat-operators-stwns\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.619980 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.683423 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvsh5"] Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.695598 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22dv4\" (UniqueName: \"kubernetes.io/projected/66135764-cd30-4365-bc33-c1fb96eedc85-kube-api-access-22dv4\") pod \"redhat-operators-stwns\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.695695 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-catalog-content\") pod \"redhat-operators-stwns\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.695783 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-utilities\") pod \"redhat-operators-stwns\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.696380 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-catalog-content\") pod \"redhat-operators-stwns\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.696504 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-utilities\") pod \"redhat-operators-stwns\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.710387 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.720811 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22dv4\" (UniqueName: \"kubernetes.io/projected/66135764-cd30-4365-bc33-c1fb96eedc85-kube-api-access-22dv4\") pod \"redhat-operators-stwns\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.737945 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nmsxl"] Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.829224 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.900323 4923 generic.go:334] "Generic (PLEG): container finished" podID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerID="83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8" exitCode=0 Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.900467 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vs8dt" event={"ID":"09e8da7b-adaf-49f9-9811-9ca90e2bff56","Type":"ContainerDied","Data":"83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8"} Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.902634 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" event={"ID":"02a21759-421a-429b-ac1b-c43cd7c236f0","Type":"ContainerStarted","Data":"6bb80ea12b6e32d97fd4826c6900da6eae9909aa253d2360d04452fc9523f4ee"} Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.911514 4923 generic.go:334] "Generic (PLEG): container finished" podID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerID="733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb" exitCode=0 Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.911940 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkxzd" event={"ID":"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95","Type":"ContainerDied","Data":"733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb"} Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.912034 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkxzd" event={"ID":"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95","Type":"ContainerStarted","Data":"b410f7087b53f5548f197dc6bfb18cb6ba64e0ecd6aba5380e9da42b2003cb94"} Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.920158 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmsxl" event={"ID":"fa810245-ceec-4d06-b02d-20e39d65ff1f","Type":"ContainerStarted","Data":"8d47c8fcd18bc8bf033991c89b78cb3c68ab0b886100b0714be53445199b5583"} Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.931120 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-p9k66" event={"ID":"eb5ecfa1-07aa-48ca-8727-a28a935fd937","Type":"ContainerStarted","Data":"85ab6390a2b8f04ab3fe134de9bc6441ff6c56cc305d5becf846dbafb027c655"} Oct 09 10:07:47 crc kubenswrapper[4923]: I1009 10:07:47.945530 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-gv7qk" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.010592 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.047871 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-p9k66" podStartSLOduration=14.047838589 podStartE2EDuration="14.047838589s" podCreationTimestamp="2025-10-09 10:07:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:48.043613235 +0000 UTC m=+154.111794991" watchObservedRunningTime="2025-10-09 10:07:48.047838589 +0000 UTC m=+154.116020345" Oct 09 10:07:48 crc kubenswrapper[4923]: W1009 10:07:48.094681 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podef2f85bb_dede_4bd6_993f_cf7e80a9227f.slice/crio-31d5d2eb2c66e2534376281a59b73313616522d456e5d8bf1098a4ab8f0a1816 WatchSource:0}: Error finding container 31d5d2eb2c66e2534376281a59b73313616522d456e5d8bf1098a4ab8f0a1816: Status 404 returned error can't find the container with id 31d5d2eb2c66e2534376281a59b73313616522d456e5d8bf1098a4ab8f0a1816 Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.372103 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:48 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:48 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:48 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.372605 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.408001 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.451434 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-stwns"] Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.509422 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/215543fc-99e1-47f3-9ffd-9adb4fd9c611-secret-volume\") pod \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.509520 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs9tk\" (UniqueName: \"kubernetes.io/projected/215543fc-99e1-47f3-9ffd-9adb4fd9c611-kube-api-access-rs9tk\") pod \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.509541 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/215543fc-99e1-47f3-9ffd-9adb4fd9c611-config-volume\") pod \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\" (UID: \"215543fc-99e1-47f3-9ffd-9adb4fd9c611\") " Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.511815 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/215543fc-99e1-47f3-9ffd-9adb4fd9c611-config-volume" (OuterVolumeSpecName: "config-volume") pod "215543fc-99e1-47f3-9ffd-9adb4fd9c611" (UID: "215543fc-99e1-47f3-9ffd-9adb4fd9c611"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:07:48 crc kubenswrapper[4923]: W1009 10:07:48.517985 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66135764_cd30_4365_bc33_c1fb96eedc85.slice/crio-de450717c0526a377102f99d2c457aed74a221015a9d2f147a986537f8c5adfd WatchSource:0}: Error finding container de450717c0526a377102f99d2c457aed74a221015a9d2f147a986537f8c5adfd: Status 404 returned error can't find the container with id de450717c0526a377102f99d2c457aed74a221015a9d2f147a986537f8c5adfd Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.519644 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/215543fc-99e1-47f3-9ffd-9adb4fd9c611-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "215543fc-99e1-47f3-9ffd-9adb4fd9c611" (UID: "215543fc-99e1-47f3-9ffd-9adb4fd9c611"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.533187 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/215543fc-99e1-47f3-9ffd-9adb4fd9c611-kube-api-access-rs9tk" (OuterVolumeSpecName: "kube-api-access-rs9tk") pod "215543fc-99e1-47f3-9ffd-9adb4fd9c611" (UID: "215543fc-99e1-47f3-9ffd-9adb4fd9c611"). InnerVolumeSpecName "kube-api-access-rs9tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.611829 4923 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/215543fc-99e1-47f3-9ffd-9adb4fd9c611-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.611868 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs9tk\" (UniqueName: \"kubernetes.io/projected/215543fc-99e1-47f3-9ffd-9adb4fd9c611-kube-api-access-rs9tk\") on node \"crc\" DevicePath \"\"" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.611878 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/215543fc-99e1-47f3-9ffd-9adb4fd9c611-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.626525 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.940808 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" event={"ID":"215543fc-99e1-47f3-9ffd-9adb4fd9c611","Type":"ContainerDied","Data":"02cdf5e5e12375439ebc594bb8bb081f9d1e10c5e1e4d28d2eed9a3c2811d3c3"} Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.940892 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02cdf5e5e12375439ebc594bb8bb081f9d1e10c5e1e4d28d2eed9a3c2811d3c3" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.940847 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.942823 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef2f85bb-dede-4bd6-993f-cf7e80a9227f","Type":"ContainerStarted","Data":"31d5d2eb2c66e2534376281a59b73313616522d456e5d8bf1098a4ab8f0a1816"} Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.964108 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" event={"ID":"02a21759-421a-429b-ac1b-c43cd7c236f0","Type":"ContainerStarted","Data":"8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407"} Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.964197 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.981001 4923 generic.go:334] "Generic (PLEG): container finished" podID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerID="074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35" exitCode=0 Oct 09 10:07:48 crc kubenswrapper[4923]: I1009 10:07:48.981271 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmsxl" event={"ID":"fa810245-ceec-4d06-b02d-20e39d65ff1f","Type":"ContainerDied","Data":"074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35"} Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.005025 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" podStartSLOduration=133.005006353 podStartE2EDuration="2m13.005006353s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:49.000274765 +0000 UTC m=+155.068456541" watchObservedRunningTime="2025-10-09 10:07:49.005006353 +0000 UTC m=+155.073188109" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.005781 4923 generic.go:334] "Generic (PLEG): container finished" podID="66135764-cd30-4365-bc33-c1fb96eedc85" containerID="33cbef3b4548776387c8ccdc5c87ddba8379c106caa71b167697df92aad32e85" exitCode=0 Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.006725 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stwns" event={"ID":"66135764-cd30-4365-bc33-c1fb96eedc85","Type":"ContainerDied","Data":"33cbef3b4548776387c8ccdc5c87ddba8379c106caa71b167697df92aad32e85"} Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.006784 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stwns" event={"ID":"66135764-cd30-4365-bc33-c1fb96eedc85","Type":"ContainerStarted","Data":"de450717c0526a377102f99d2c457aed74a221015a9d2f147a986537f8c5adfd"} Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.374689 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:49 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:49 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:49 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.375268 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.582345 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 09 10:07:49 crc kubenswrapper[4923]: E1009 10:07:49.582666 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="215543fc-99e1-47f3-9ffd-9adb4fd9c611" containerName="collect-profiles" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.582692 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="215543fc-99e1-47f3-9ffd-9adb4fd9c611" containerName="collect-profiles" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.582893 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="215543fc-99e1-47f3-9ffd-9adb4fd9c611" containerName="collect-profiles" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.583369 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.586412 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.586641 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.594391 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.741445 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a72feac0-410a-4cef-9d6c-f70bd8988482-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a72feac0-410a-4cef-9d6c-f70bd8988482\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.741508 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a72feac0-410a-4cef-9d6c-f70bd8988482-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a72feac0-410a-4cef-9d6c-f70bd8988482\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.843627 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a72feac0-410a-4cef-9d6c-f70bd8988482-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a72feac0-410a-4cef-9d6c-f70bd8988482\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.843819 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a72feac0-410a-4cef-9d6c-f70bd8988482-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a72feac0-410a-4cef-9d6c-f70bd8988482\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.844196 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a72feac0-410a-4cef-9d6c-f70bd8988482-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"a72feac0-410a-4cef-9d6c-f70bd8988482\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.867417 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a72feac0-410a-4cef-9d6c-f70bd8988482-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"a72feac0-410a-4cef-9d6c-f70bd8988482\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:49 crc kubenswrapper[4923]: I1009 10:07:49.905820 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:50 crc kubenswrapper[4923]: I1009 10:07:50.022029 4923 generic.go:334] "Generic (PLEG): container finished" podID="ef2f85bb-dede-4bd6-993f-cf7e80a9227f" containerID="fb2e72455f4295d219aa11b374d53aec47441b715e142f328d1c07f914a5c40d" exitCode=0 Oct 09 10:07:50 crc kubenswrapper[4923]: I1009 10:07:50.022140 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef2f85bb-dede-4bd6-993f-cf7e80a9227f","Type":"ContainerDied","Data":"fb2e72455f4295d219aa11b374d53aec47441b715e142f328d1c07f914a5c40d"} Oct 09 10:07:50 crc kubenswrapper[4923]: I1009 10:07:50.250499 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 09 10:07:50 crc kubenswrapper[4923]: I1009 10:07:50.367127 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:50 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:50 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:50 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:50 crc kubenswrapper[4923]: I1009 10:07:50.367219 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.037448 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a72feac0-410a-4cef-9d6c-f70bd8988482","Type":"ContainerStarted","Data":"8ed8017e709e01a76efa9b732cf9508a6d221d52346642dedde5d065b1973726"} Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.366262 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:51 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:51 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:51 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.366348 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.413663 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.582530 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kubelet-dir\") pod \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\" (UID: \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\") " Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.582677 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kube-api-access\") pod \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\" (UID: \"ef2f85bb-dede-4bd6-993f-cf7e80a9227f\") " Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.582703 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "ef2f85bb-dede-4bd6-993f-cf7e80a9227f" (UID: "ef2f85bb-dede-4bd6-993f-cf7e80a9227f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.583102 4923 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.589124 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "ef2f85bb-dede-4bd6-993f-cf7e80a9227f" (UID: "ef2f85bb-dede-4bd6-993f-cf7e80a9227f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.684983 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ef2f85bb-dede-4bd6-993f-cf7e80a9227f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.953896 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:51 crc kubenswrapper[4923]: I1009 10:07:51.960728 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-dtw6g" Oct 09 10:07:52 crc kubenswrapper[4923]: I1009 10:07:52.049619 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a72feac0-410a-4cef-9d6c-f70bd8988482","Type":"ContainerStarted","Data":"70ac97dfb3fcd5e7388a6acf0cc548cfef596f3ca073c73e4804f371524b12a2"} Oct 09 10:07:52 crc kubenswrapper[4923]: I1009 10:07:52.052321 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 10:07:52 crc kubenswrapper[4923]: I1009 10:07:52.057963 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"ef2f85bb-dede-4bd6-993f-cf7e80a9227f","Type":"ContainerDied","Data":"31d5d2eb2c66e2534376281a59b73313616522d456e5d8bf1098a4ab8f0a1816"} Oct 09 10:07:52 crc kubenswrapper[4923]: I1009 10:07:52.058142 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31d5d2eb2c66e2534376281a59b73313616522d456e5d8bf1098a4ab8f0a1816" Oct 09 10:07:52 crc kubenswrapper[4923]: I1009 10:07:52.169590 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.169538181 podStartE2EDuration="3.169538181s" podCreationTimestamp="2025-10-09 10:07:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:07:52.134493726 +0000 UTC m=+158.202675482" watchObservedRunningTime="2025-10-09 10:07:52.169538181 +0000 UTC m=+158.237719937" Oct 09 10:07:52 crc kubenswrapper[4923]: I1009 10:07:52.366298 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:52 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:52 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:52 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:52 crc kubenswrapper[4923]: I1009 10:07:52.366369 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:53 crc kubenswrapper[4923]: I1009 10:07:53.078556 4923 generic.go:334] "Generic (PLEG): container finished" podID="a72feac0-410a-4cef-9d6c-f70bd8988482" containerID="70ac97dfb3fcd5e7388a6acf0cc548cfef596f3ca073c73e4804f371524b12a2" exitCode=0 Oct 09 10:07:53 crc kubenswrapper[4923]: I1009 10:07:53.078609 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a72feac0-410a-4cef-9d6c-f70bd8988482","Type":"ContainerDied","Data":"70ac97dfb3fcd5e7388a6acf0cc548cfef596f3ca073c73e4804f371524b12a2"} Oct 09 10:07:53 crc kubenswrapper[4923]: I1009 10:07:53.144610 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-8fckr" Oct 09 10:07:53 crc kubenswrapper[4923]: I1009 10:07:53.365414 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:53 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:53 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:53 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:53 crc kubenswrapper[4923]: I1009 10:07:53.365860 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.367419 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:54 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:54 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:54 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.367492 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.423635 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.563962 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a72feac0-410a-4cef-9d6c-f70bd8988482-kube-api-access\") pod \"a72feac0-410a-4cef-9d6c-f70bd8988482\" (UID: \"a72feac0-410a-4cef-9d6c-f70bd8988482\") " Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.564074 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a72feac0-410a-4cef-9d6c-f70bd8988482-kubelet-dir\") pod \"a72feac0-410a-4cef-9d6c-f70bd8988482\" (UID: \"a72feac0-410a-4cef-9d6c-f70bd8988482\") " Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.564321 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a72feac0-410a-4cef-9d6c-f70bd8988482-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a72feac0-410a-4cef-9d6c-f70bd8988482" (UID: "a72feac0-410a-4cef-9d6c-f70bd8988482"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.564762 4923 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a72feac0-410a-4cef-9d6c-f70bd8988482-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.573948 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a72feac0-410a-4cef-9d6c-f70bd8988482-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a72feac0-410a-4cef-9d6c-f70bd8988482" (UID: "a72feac0-410a-4cef-9d6c-f70bd8988482"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.600548 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.600624 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:07:54 crc kubenswrapper[4923]: I1009 10:07:54.666403 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a72feac0-410a-4cef-9d6c-f70bd8988482-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 10:07:55 crc kubenswrapper[4923]: I1009 10:07:55.129140 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"a72feac0-410a-4cef-9d6c-f70bd8988482","Type":"ContainerDied","Data":"8ed8017e709e01a76efa9b732cf9508a6d221d52346642dedde5d065b1973726"} Oct 09 10:07:55 crc kubenswrapper[4923]: I1009 10:07:55.129240 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ed8017e709e01a76efa9b732cf9508a6d221d52346642dedde5d065b1973726" Oct 09 10:07:55 crc kubenswrapper[4923]: I1009 10:07:55.129421 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 10:07:55 crc kubenswrapper[4923]: I1009 10:07:55.364475 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:55 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:55 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:55 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:55 crc kubenswrapper[4923]: I1009 10:07:55.365024 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:56 crc kubenswrapper[4923]: I1009 10:07:56.365493 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:56 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:56 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:56 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:56 crc kubenswrapper[4923]: I1009 10:07:56.365552 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:56 crc kubenswrapper[4923]: I1009 10:07:56.490561 4923 patch_prober.go:28] interesting pod/console-f9d7485db-66f2z container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Oct 09 10:07:56 crc kubenswrapper[4923]: I1009 10:07:56.490615 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-66f2z" podUID="a8c487fa-3436-4cc9-83b3-de49256e97c9" containerName="console" probeResult="failure" output="Get \"https://10.217.0.7:8443/health\": dial tcp 10.217.0.7:8443: connect: connection refused" Oct 09 10:07:57 crc kubenswrapper[4923]: I1009 10:07:57.364548 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:57 crc kubenswrapper[4923]: [-]has-synced failed: reason withheld Oct 09 10:07:57 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:57 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:57 crc kubenswrapper[4923]: I1009 10:07:57.364625 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:57 crc kubenswrapper[4923]: I1009 10:07:57.436876 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-9rq8l container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 09 10:07:57 crc kubenswrapper[4923]: I1009 10:07:57.436936 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-9rq8l" podUID="dc7bae79-3278-46f0-91fd-9a49b4057200" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 09 10:07:57 crc kubenswrapper[4923]: I1009 10:07:57.437069 4923 patch_prober.go:28] interesting pod/downloads-7954f5f757-9rq8l container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Oct 09 10:07:57 crc kubenswrapper[4923]: I1009 10:07:57.437134 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-9rq8l" podUID="dc7bae79-3278-46f0-91fd-9a49b4057200" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Oct 09 10:07:58 crc kubenswrapper[4923]: I1009 10:07:58.365309 4923 patch_prober.go:28] interesting pod/router-default-5444994796-85h6b container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 10:07:58 crc kubenswrapper[4923]: [+]has-synced ok Oct 09 10:07:58 crc kubenswrapper[4923]: [+]process-running ok Oct 09 10:07:58 crc kubenswrapper[4923]: healthz check failed Oct 09 10:07:58 crc kubenswrapper[4923]: I1009 10:07:58.365685 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-85h6b" podUID="68e9fb5c-167b-4528-9c31-97cbdc41aecd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:07:58 crc kubenswrapper[4923]: I1009 10:07:58.742895 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:58 crc kubenswrapper[4923]: I1009 10:07:58.750587 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0ffdd04f-94e8-4fca-9676-ef87952debcc-metrics-certs\") pod \"network-metrics-daemon-dxnsb\" (UID: \"0ffdd04f-94e8-4fca-9676-ef87952debcc\") " pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:58 crc kubenswrapper[4923]: I1009 10:07:58.935979 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-dxnsb" Oct 09 10:07:59 crc kubenswrapper[4923]: I1009 10:07:59.366387 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:07:59 crc kubenswrapper[4923]: I1009 10:07:59.369799 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-85h6b" Oct 09 10:08:06 crc kubenswrapper[4923]: I1009 10:08:06.496491 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:08:06 crc kubenswrapper[4923]: I1009 10:08:06.501189 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:08:07 crc kubenswrapper[4923]: I1009 10:08:07.339359 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:08:07 crc kubenswrapper[4923]: I1009 10:08:07.460127 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-9rq8l" Oct 09 10:08:17 crc kubenswrapper[4923]: I1009 10:08:17.805073 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-b9kd6" Oct 09 10:08:20 crc kubenswrapper[4923]: E1009 10:08:20.991734 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 09 10:08:20 crc kubenswrapper[4923]: E1009 10:08:20.992484 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-22dv4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-stwns_openshift-marketplace(66135764-cd30-4365-bc33-c1fb96eedc85): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 10:08:20 crc kubenswrapper[4923]: E1009 10:08:20.993730 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-stwns" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" Oct 09 10:08:21 crc kubenswrapper[4923]: E1009 10:08:21.368806 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-stwns" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" Oct 09 10:08:23 crc kubenswrapper[4923]: I1009 10:08:23.659174 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 10:08:24 crc kubenswrapper[4923]: I1009 10:08:24.599979 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:08:24 crc kubenswrapper[4923]: I1009 10:08:24.600076 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:08:26 crc kubenswrapper[4923]: E1009 10:08:26.427430 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 09 10:08:26 crc kubenswrapper[4923]: E1009 10:08:26.427600 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qmhct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-r7pbq_openshift-marketplace(39ceb27f-3ad7-452a-bbdd-92057099d011): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 10:08:26 crc kubenswrapper[4923]: E1009 10:08:26.428907 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-r7pbq" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.406819 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-r7pbq" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.489196 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.489556 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kl8tr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-b8fd8_openshift-marketplace(d4a3f04f-d392-4065-896b-7db7683f50e6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.490649 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-b8fd8" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.766477 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.767098 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xkwhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-jbd8g_openshift-marketplace(31d5c282-dd47-420c-bc6e-cfcbee938bfc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.769336 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-jbd8g" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.800514 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.800668 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mp2jb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-qhv8h_openshift-marketplace(e46ac356-170c-4c91-b8b6-36ec0dfb93d8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.802016 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-qhv8h" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.858627 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.858799 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p4pj8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-zkxzd_openshift-marketplace(8c0567db-2f9d-46cf-bfc8-b3893c2a4c95): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 10:08:36 crc kubenswrapper[4923]: E1009 10:08:36.859946 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-zkxzd" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" Oct 09 10:08:37 crc kubenswrapper[4923]: I1009 10:08:37.027577 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-dxnsb"] Oct 09 10:08:37 crc kubenswrapper[4923]: W1009 10:08:37.059351 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ffdd04f_94e8_4fca_9676_ef87952debcc.slice/crio-ec34efd115e1aab3c35408b89686d39c789dc9b0c0c245144d687f88d21f805d WatchSource:0}: Error finding container ec34efd115e1aab3c35408b89686d39c789dc9b0c0c245144d687f88d21f805d: Status 404 returned error can't find the container with id ec34efd115e1aab3c35408b89686d39c789dc9b0c0c245144d687f88d21f805d Oct 09 10:08:37 crc kubenswrapper[4923]: I1009 10:08:37.409882 4923 generic.go:334] "Generic (PLEG): container finished" podID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerID="2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f" exitCode=0 Oct 09 10:08:37 crc kubenswrapper[4923]: I1009 10:08:37.409960 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vs8dt" event={"ID":"09e8da7b-adaf-49f9-9811-9ca90e2bff56","Type":"ContainerDied","Data":"2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f"} Oct 09 10:08:37 crc kubenswrapper[4923]: I1009 10:08:37.415117 4923 generic.go:334] "Generic (PLEG): container finished" podID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerID="a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10" exitCode=0 Oct 09 10:08:37 crc kubenswrapper[4923]: I1009 10:08:37.415185 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmsxl" event={"ID":"fa810245-ceec-4d06-b02d-20e39d65ff1f","Type":"ContainerDied","Data":"a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10"} Oct 09 10:08:37 crc kubenswrapper[4923]: I1009 10:08:37.417727 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" event={"ID":"0ffdd04f-94e8-4fca-9676-ef87952debcc","Type":"ContainerStarted","Data":"9dd3909494b516a2a943cdf2d7de7153c2d353742383a5e236e76cd4add77889"} Oct 09 10:08:37 crc kubenswrapper[4923]: I1009 10:08:37.417813 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" event={"ID":"0ffdd04f-94e8-4fca-9676-ef87952debcc","Type":"ContainerStarted","Data":"ec34efd115e1aab3c35408b89686d39c789dc9b0c0c245144d687f88d21f805d"} Oct 09 10:08:37 crc kubenswrapper[4923]: E1009 10:08:37.420235 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-jbd8g" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" Oct 09 10:08:37 crc kubenswrapper[4923]: E1009 10:08:37.420245 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-qhv8h" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" Oct 09 10:08:37 crc kubenswrapper[4923]: E1009 10:08:37.427999 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-b8fd8" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" Oct 09 10:08:37 crc kubenswrapper[4923]: E1009 10:08:37.432336 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-zkxzd" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" Oct 09 10:08:38 crc kubenswrapper[4923]: I1009 10:08:38.424744 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vs8dt" event={"ID":"09e8da7b-adaf-49f9-9811-9ca90e2bff56","Type":"ContainerStarted","Data":"bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1"} Oct 09 10:08:38 crc kubenswrapper[4923]: I1009 10:08:38.427350 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmsxl" event={"ID":"fa810245-ceec-4d06-b02d-20e39d65ff1f","Type":"ContainerStarted","Data":"c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533"} Oct 09 10:08:38 crc kubenswrapper[4923]: I1009 10:08:38.429885 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-dxnsb" event={"ID":"0ffdd04f-94e8-4fca-9676-ef87952debcc","Type":"ContainerStarted","Data":"129807f7f6bdbc8c54ac71972ec71577cc2689da9c6d8f12e609fb76564a3e8b"} Oct 09 10:08:38 crc kubenswrapper[4923]: I1009 10:08:38.435071 4923 generic.go:334] "Generic (PLEG): container finished" podID="66135764-cd30-4365-bc33-c1fb96eedc85" containerID="070a2b04d278a19ddcb49047fa2fa45e9131fbeb933b5c4baa583cda4e543111" exitCode=0 Oct 09 10:08:38 crc kubenswrapper[4923]: I1009 10:08:38.435124 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stwns" event={"ID":"66135764-cd30-4365-bc33-c1fb96eedc85","Type":"ContainerDied","Data":"070a2b04d278a19ddcb49047fa2fa45e9131fbeb933b5c4baa583cda4e543111"} Oct 09 10:08:38 crc kubenswrapper[4923]: I1009 10:08:38.445530 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vs8dt" podStartSLOduration=3.608643238 podStartE2EDuration="53.44550951s" podCreationTimestamp="2025-10-09 10:07:45 +0000 UTC" firstStartedPulling="2025-10-09 10:07:48.030741227 +0000 UTC m=+154.098922983" lastFinishedPulling="2025-10-09 10:08:37.867607489 +0000 UTC m=+203.935789255" observedRunningTime="2025-10-09 10:08:38.44514941 +0000 UTC m=+204.513331186" watchObservedRunningTime="2025-10-09 10:08:38.44550951 +0000 UTC m=+204.513691266" Oct 09 10:08:38 crc kubenswrapper[4923]: I1009 10:08:38.480952 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nmsxl" podStartSLOduration=2.300117216 podStartE2EDuration="51.480934002s" podCreationTimestamp="2025-10-09 10:07:47 +0000 UTC" firstStartedPulling="2025-10-09 10:07:48.995212398 +0000 UTC m=+155.063394154" lastFinishedPulling="2025-10-09 10:08:38.176029184 +0000 UTC m=+204.244210940" observedRunningTime="2025-10-09 10:08:38.48048025 +0000 UTC m=+204.548662006" watchObservedRunningTime="2025-10-09 10:08:38.480934002 +0000 UTC m=+204.549115758" Oct 09 10:08:40 crc kubenswrapper[4923]: I1009 10:08:40.448590 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stwns" event={"ID":"66135764-cd30-4365-bc33-c1fb96eedc85","Type":"ContainerStarted","Data":"880d6bb973a175541aa5f5032044d55ff94c6d4d38738f00d4d1080b986f3bb7"} Oct 09 10:08:40 crc kubenswrapper[4923]: I1009 10:08:40.466801 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-dxnsb" podStartSLOduration=184.466783154 podStartE2EDuration="3m4.466783154s" podCreationTimestamp="2025-10-09 10:05:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:08:38.497304006 +0000 UTC m=+204.565485772" watchObservedRunningTime="2025-10-09 10:08:40.466783154 +0000 UTC m=+206.534964910" Oct 09 10:08:40 crc kubenswrapper[4923]: I1009 10:08:40.467967 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-stwns" podStartSLOduration=3.050966219 podStartE2EDuration="53.467960826s" podCreationTimestamp="2025-10-09 10:07:47 +0000 UTC" firstStartedPulling="2025-10-09 10:07:49.008902147 +0000 UTC m=+155.077083903" lastFinishedPulling="2025-10-09 10:08:39.425896754 +0000 UTC m=+205.494078510" observedRunningTime="2025-10-09 10:08:40.465157259 +0000 UTC m=+206.533339015" watchObservedRunningTime="2025-10-09 10:08:40.467960826 +0000 UTC m=+206.536142582" Oct 09 10:08:46 crc kubenswrapper[4923]: I1009 10:08:46.188454 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:08:46 crc kubenswrapper[4923]: I1009 10:08:46.189149 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:08:46 crc kubenswrapper[4923]: I1009 10:08:46.503806 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:08:46 crc kubenswrapper[4923]: I1009 10:08:46.543932 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:08:47 crc kubenswrapper[4923]: I1009 10:08:47.413444 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:08:47 crc kubenswrapper[4923]: I1009 10:08:47.413497 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:08:47 crc kubenswrapper[4923]: I1009 10:08:47.463039 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:08:47 crc kubenswrapper[4923]: I1009 10:08:47.527741 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:08:47 crc kubenswrapper[4923]: I1009 10:08:47.830337 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:08:47 crc kubenswrapper[4923]: I1009 10:08:47.830741 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:08:47 crc kubenswrapper[4923]: I1009 10:08:47.874719 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:08:48 crc kubenswrapper[4923]: I1009 10:08:48.537796 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:08:50 crc kubenswrapper[4923]: I1009 10:08:50.840239 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-stwns"] Oct 09 10:08:51 crc kubenswrapper[4923]: I1009 10:08:51.510225 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-stwns" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" containerName="registry-server" containerID="cri-o://880d6bb973a175541aa5f5032044d55ff94c6d4d38738f00d4d1080b986f3bb7" gracePeriod=2 Oct 09 10:08:52 crc kubenswrapper[4923]: I1009 10:08:52.518922 4923 generic.go:334] "Generic (PLEG): container finished" podID="66135764-cd30-4365-bc33-c1fb96eedc85" containerID="880d6bb973a175541aa5f5032044d55ff94c6d4d38738f00d4d1080b986f3bb7" exitCode=0 Oct 09 10:08:52 crc kubenswrapper[4923]: I1009 10:08:52.518977 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stwns" event={"ID":"66135764-cd30-4365-bc33-c1fb96eedc85","Type":"ContainerDied","Data":"880d6bb973a175541aa5f5032044d55ff94c6d4d38738f00d4d1080b986f3bb7"} Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.599715 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.599994 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.600035 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.600471 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.600561 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0" gracePeriod=600 Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.776254 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.818188 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-utilities\") pod \"66135764-cd30-4365-bc33-c1fb96eedc85\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.819174 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-utilities" (OuterVolumeSpecName: "utilities") pod "66135764-cd30-4365-bc33-c1fb96eedc85" (UID: "66135764-cd30-4365-bc33-c1fb96eedc85"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.819354 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22dv4\" (UniqueName: \"kubernetes.io/projected/66135764-cd30-4365-bc33-c1fb96eedc85-kube-api-access-22dv4\") pod \"66135764-cd30-4365-bc33-c1fb96eedc85\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.819566 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-catalog-content\") pod \"66135764-cd30-4365-bc33-c1fb96eedc85\" (UID: \"66135764-cd30-4365-bc33-c1fb96eedc85\") " Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.819918 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.837162 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66135764-cd30-4365-bc33-c1fb96eedc85-kube-api-access-22dv4" (OuterVolumeSpecName: "kube-api-access-22dv4") pod "66135764-cd30-4365-bc33-c1fb96eedc85" (UID: "66135764-cd30-4365-bc33-c1fb96eedc85"). InnerVolumeSpecName "kube-api-access-22dv4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.922153 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22dv4\" (UniqueName: \"kubernetes.io/projected/66135764-cd30-4365-bc33-c1fb96eedc85-kube-api-access-22dv4\") on node \"crc\" DevicePath \"\"" Oct 09 10:08:54 crc kubenswrapper[4923]: I1009 10:08:54.930100 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "66135764-cd30-4365-bc33-c1fb96eedc85" (UID: "66135764-cd30-4365-bc33-c1fb96eedc85"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.023447 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/66135764-cd30-4365-bc33-c1fb96eedc85-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.539402 4923 generic.go:334] "Generic (PLEG): container finished" podID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerID="14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d" exitCode=0 Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.539486 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhv8h" event={"ID":"e46ac356-170c-4c91-b8b6-36ec0dfb93d8","Type":"ContainerDied","Data":"14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d"} Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.545576 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbd8g" event={"ID":"31d5c282-dd47-420c-bc6e-cfcbee938bfc","Type":"ContainerStarted","Data":"7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8"} Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.548584 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7pbq" event={"ID":"39ceb27f-3ad7-452a-bbdd-92057099d011","Type":"ContainerStarted","Data":"50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b"} Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.551034 4923 generic.go:334] "Generic (PLEG): container finished" podID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerID="de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1" exitCode=0 Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.551099 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkxzd" event={"ID":"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95","Type":"ContainerDied","Data":"de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1"} Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.553875 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8fd8" event={"ID":"d4a3f04f-d392-4065-896b-7db7683f50e6","Type":"ContainerStarted","Data":"a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e"} Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.558646 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stwns" event={"ID":"66135764-cd30-4365-bc33-c1fb96eedc85","Type":"ContainerDied","Data":"de450717c0526a377102f99d2c457aed74a221015a9d2f147a986537f8c5adfd"} Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.558712 4923 scope.go:117] "RemoveContainer" containerID="880d6bb973a175541aa5f5032044d55ff94c6d4d38738f00d4d1080b986f3bb7" Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.558744 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stwns" Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.564498 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0" exitCode=0 Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.564589 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0"} Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.564638 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"fdc7d576a4e21a1ee542a4eba76ea241d8e56b0e66e0f288027bd3d2de345b25"} Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.769434 4923 scope.go:117] "RemoveContainer" containerID="070a2b04d278a19ddcb49047fa2fa45e9131fbeb933b5c4baa583cda4e543111" Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.785321 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-stwns"] Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.795273 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-stwns"] Oct 09 10:08:55 crc kubenswrapper[4923]: I1009 10:08:55.806941 4923 scope.go:117] "RemoveContainer" containerID="33cbef3b4548776387c8ccdc5c87ddba8379c106caa71b167697df92aad32e85" Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.574152 4923 generic.go:334] "Generic (PLEG): container finished" podID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerID="7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8" exitCode=0 Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.574243 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbd8g" event={"ID":"31d5c282-dd47-420c-bc6e-cfcbee938bfc","Type":"ContainerDied","Data":"7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8"} Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.578884 4923 generic.go:334] "Generic (PLEG): container finished" podID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerID="50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b" exitCode=0 Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.578981 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7pbq" event={"ID":"39ceb27f-3ad7-452a-bbdd-92057099d011","Type":"ContainerDied","Data":"50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b"} Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.581338 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkxzd" event={"ID":"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95","Type":"ContainerStarted","Data":"120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91"} Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.584154 4923 generic.go:334] "Generic (PLEG): container finished" podID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerID="a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e" exitCode=0 Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.584218 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8fd8" event={"ID":"d4a3f04f-d392-4065-896b-7db7683f50e6","Type":"ContainerDied","Data":"a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e"} Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.614283 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" path="/var/lib/kubelet/pods/66135764-cd30-4365-bc33-c1fb96eedc85/volumes" Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.615210 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.615266 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:08:56 crc kubenswrapper[4923]: I1009 10:08:56.632906 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zkxzd" podStartSLOduration=2.3708967530000002 podStartE2EDuration="1m10.632883069s" podCreationTimestamp="2025-10-09 10:07:46 +0000 UTC" firstStartedPulling="2025-10-09 10:07:48.031238581 +0000 UTC m=+154.099420337" lastFinishedPulling="2025-10-09 10:08:56.293224897 +0000 UTC m=+222.361406653" observedRunningTime="2025-10-09 10:08:56.629563387 +0000 UTC m=+222.697745143" watchObservedRunningTime="2025-10-09 10:08:56.632883069 +0000 UTC m=+222.701064825" Oct 09 10:08:57 crc kubenswrapper[4923]: I1009 10:08:57.597518 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhv8h" event={"ID":"e46ac356-170c-4c91-b8b6-36ec0dfb93d8","Type":"ContainerStarted","Data":"9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0"} Oct 09 10:08:57 crc kubenswrapper[4923]: I1009 10:08:57.600409 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbd8g" event={"ID":"31d5c282-dd47-420c-bc6e-cfcbee938bfc","Type":"ContainerStarted","Data":"04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042"} Oct 09 10:08:57 crc kubenswrapper[4923]: I1009 10:08:57.602662 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7pbq" event={"ID":"39ceb27f-3ad7-452a-bbdd-92057099d011","Type":"ContainerStarted","Data":"4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9"} Oct 09 10:08:57 crc kubenswrapper[4923]: I1009 10:08:57.605243 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8fd8" event={"ID":"d4a3f04f-d392-4065-896b-7db7683f50e6","Type":"ContainerStarted","Data":"427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51"} Oct 09 10:08:57 crc kubenswrapper[4923]: I1009 10:08:57.659862 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r7pbq" podStartSLOduration=4.47574789 podStartE2EDuration="1m14.659809301s" podCreationTimestamp="2025-10-09 10:07:43 +0000 UTC" firstStartedPulling="2025-10-09 10:07:46.84593597 +0000 UTC m=+152.914117726" lastFinishedPulling="2025-10-09 10:08:57.029997381 +0000 UTC m=+223.098179137" observedRunningTime="2025-10-09 10:08:57.653189808 +0000 UTC m=+223.721371584" watchObservedRunningTime="2025-10-09 10:08:57.659809301 +0000 UTC m=+223.727991057" Oct 09 10:08:57 crc kubenswrapper[4923]: I1009 10:08:57.660948 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-zkxzd" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerName="registry-server" probeResult="failure" output=< Oct 09 10:08:57 crc kubenswrapper[4923]: timeout: failed to connect service ":50051" within 1s Oct 09 10:08:57 crc kubenswrapper[4923]: > Oct 09 10:08:57 crc kubenswrapper[4923]: I1009 10:08:57.661744 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-qhv8h" podStartSLOduration=4.043979713 podStartE2EDuration="1m13.661733085s" podCreationTimestamp="2025-10-09 10:07:44 +0000 UTC" firstStartedPulling="2025-10-09 10:07:46.862325473 +0000 UTC m=+152.930507229" lastFinishedPulling="2025-10-09 10:08:56.480078845 +0000 UTC m=+222.548260601" observedRunningTime="2025-10-09 10:08:57.626032245 +0000 UTC m=+223.694214001" watchObservedRunningTime="2025-10-09 10:08:57.661733085 +0000 UTC m=+223.729914841" Oct 09 10:08:57 crc kubenswrapper[4923]: I1009 10:08:57.680810 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jbd8g" podStartSLOduration=3.410397835 podStartE2EDuration="1m13.680788193s" podCreationTimestamp="2025-10-09 10:07:44 +0000 UTC" firstStartedPulling="2025-10-09 10:07:46.881404128 +0000 UTC m=+152.949585884" lastFinishedPulling="2025-10-09 10:08:57.151794486 +0000 UTC m=+223.219976242" observedRunningTime="2025-10-09 10:08:57.67851947 +0000 UTC m=+223.746701236" watchObservedRunningTime="2025-10-09 10:08:57.680788193 +0000 UTC m=+223.748969949" Oct 09 10:08:57 crc kubenswrapper[4923]: I1009 10:08:57.704903 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b8fd8" podStartSLOduration=3.475612716 podStartE2EDuration="1m13.70488099s" podCreationTimestamp="2025-10-09 10:07:44 +0000 UTC" firstStartedPulling="2025-10-09 10:07:46.847843492 +0000 UTC m=+152.916025258" lastFinishedPulling="2025-10-09 10:08:57.077111776 +0000 UTC m=+223.145293532" observedRunningTime="2025-10-09 10:08:57.704544051 +0000 UTC m=+223.772725807" watchObservedRunningTime="2025-10-09 10:08:57.70488099 +0000 UTC m=+223.773062746" Oct 09 10:09:02 crc kubenswrapper[4923]: I1009 10:09:02.027088 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lwzgh"] Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.546457 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.546823 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.599452 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.633508 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.633547 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.673601 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.673652 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.674343 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.689182 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.725424 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.864830 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.864879 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:09:04 crc kubenswrapper[4923]: I1009 10:09:04.910072 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:09:05 crc kubenswrapper[4923]: I1009 10:09:05.690688 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:09:05 crc kubenswrapper[4923]: I1009 10:09:05.694272 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:09:05 crc kubenswrapper[4923]: I1009 10:09:05.703881 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:09:06 crc kubenswrapper[4923]: I1009 10:09:06.680876 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:09:06 crc kubenswrapper[4923]: I1009 10:09:06.731500 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:09:07 crc kubenswrapper[4923]: I1009 10:09:07.639662 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jbd8g"] Oct 09 10:09:07 crc kubenswrapper[4923]: I1009 10:09:07.662273 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jbd8g" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerName="registry-server" containerID="cri-o://04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042" gracePeriod=2 Oct 09 10:09:08 crc kubenswrapper[4923]: I1009 10:09:08.637566 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qhv8h"] Oct 09 10:09:08 crc kubenswrapper[4923]: I1009 10:09:08.638142 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-qhv8h" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerName="registry-server" containerID="cri-o://9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0" gracePeriod=2 Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.257359 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.321903 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.342651 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-catalog-content\") pod \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.342714 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-utilities\") pod \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.342806 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkwhc\" (UniqueName: \"kubernetes.io/projected/31d5c282-dd47-420c-bc6e-cfcbee938bfc-kube-api-access-xkwhc\") pod \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\" (UID: \"31d5c282-dd47-420c-bc6e-cfcbee938bfc\") " Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.343873 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-utilities" (OuterVolumeSpecName: "utilities") pod "31d5c282-dd47-420c-bc6e-cfcbee938bfc" (UID: "31d5c282-dd47-420c-bc6e-cfcbee938bfc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.349424 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d5c282-dd47-420c-bc6e-cfcbee938bfc-kube-api-access-xkwhc" (OuterVolumeSpecName: "kube-api-access-xkwhc") pod "31d5c282-dd47-420c-bc6e-cfcbee938bfc" (UID: "31d5c282-dd47-420c-bc6e-cfcbee938bfc"). InnerVolumeSpecName "kube-api-access-xkwhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.386385 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31d5c282-dd47-420c-bc6e-cfcbee938bfc" (UID: "31d5c282-dd47-420c-bc6e-cfcbee938bfc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.443719 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp2jb\" (UniqueName: \"kubernetes.io/projected/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-kube-api-access-mp2jb\") pod \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.443864 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-catalog-content\") pod \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.443948 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-utilities\") pod \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\" (UID: \"e46ac356-170c-4c91-b8b6-36ec0dfb93d8\") " Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.444244 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.444260 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31d5c282-dd47-420c-bc6e-cfcbee938bfc-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.444270 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkwhc\" (UniqueName: \"kubernetes.io/projected/31d5c282-dd47-420c-bc6e-cfcbee938bfc-kube-api-access-xkwhc\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.445097 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-utilities" (OuterVolumeSpecName: "utilities") pod "e46ac356-170c-4c91-b8b6-36ec0dfb93d8" (UID: "e46ac356-170c-4c91-b8b6-36ec0dfb93d8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.447543 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-kube-api-access-mp2jb" (OuterVolumeSpecName: "kube-api-access-mp2jb") pod "e46ac356-170c-4c91-b8b6-36ec0dfb93d8" (UID: "e46ac356-170c-4c91-b8b6-36ec0dfb93d8"). InnerVolumeSpecName "kube-api-access-mp2jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.496403 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e46ac356-170c-4c91-b8b6-36ec0dfb93d8" (UID: "e46ac356-170c-4c91-b8b6-36ec0dfb93d8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.545821 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp2jb\" (UniqueName: \"kubernetes.io/projected/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-kube-api-access-mp2jb\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.545862 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.545871 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e46ac356-170c-4c91-b8b6-36ec0dfb93d8-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.680954 4923 generic.go:334] "Generic (PLEG): container finished" podID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerID="9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0" exitCode=0 Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.681031 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhv8h" event={"ID":"e46ac356-170c-4c91-b8b6-36ec0dfb93d8","Type":"ContainerDied","Data":"9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0"} Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.681068 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-qhv8h" event={"ID":"e46ac356-170c-4c91-b8b6-36ec0dfb93d8","Type":"ContainerDied","Data":"0ad6968e5da93c1944dbe9182b3f477fb5e125ecb148ece9fae8db40d050137f"} Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.681091 4923 scope.go:117] "RemoveContainer" containerID="9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.681236 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-qhv8h" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.685406 4923 generic.go:334] "Generic (PLEG): container finished" podID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerID="04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042" exitCode=0 Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.685475 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbd8g" event={"ID":"31d5c282-dd47-420c-bc6e-cfcbee938bfc","Type":"ContainerDied","Data":"04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042"} Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.685493 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jbd8g" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.685512 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jbd8g" event={"ID":"31d5c282-dd47-420c-bc6e-cfcbee938bfc","Type":"ContainerDied","Data":"ecaafaa005b7769059cf20724c96c913706cc23a9fddeeb258922d098dce9f30"} Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.702293 4923 scope.go:117] "RemoveContainer" containerID="14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.715611 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-qhv8h"] Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.719553 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-qhv8h"] Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.726033 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jbd8g"] Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.729300 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jbd8g"] Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.772686 4923 scope.go:117] "RemoveContainer" containerID="7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.788614 4923 scope.go:117] "RemoveContainer" containerID="9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0" Oct 09 10:09:09 crc kubenswrapper[4923]: E1009 10:09:09.789462 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0\": container with ID starting with 9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0 not found: ID does not exist" containerID="9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.789509 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0"} err="failed to get container status \"9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0\": rpc error: code = NotFound desc = could not find container \"9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0\": container with ID starting with 9385787a10e677971aea504ab9486f16333f088154fa952cb47a158bee56c6b0 not found: ID does not exist" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.789540 4923 scope.go:117] "RemoveContainer" containerID="14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d" Oct 09 10:09:09 crc kubenswrapper[4923]: E1009 10:09:09.790261 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d\": container with ID starting with 14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d not found: ID does not exist" containerID="14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.790311 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d"} err="failed to get container status \"14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d\": rpc error: code = NotFound desc = could not find container \"14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d\": container with ID starting with 14bb690d34cecbc35a565c388745a7b6347d4436491714b749e35ade3ca8ec1d not found: ID does not exist" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.790353 4923 scope.go:117] "RemoveContainer" containerID="7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05" Oct 09 10:09:09 crc kubenswrapper[4923]: E1009 10:09:09.790966 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05\": container with ID starting with 7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05 not found: ID does not exist" containerID="7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.791008 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05"} err="failed to get container status \"7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05\": rpc error: code = NotFound desc = could not find container \"7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05\": container with ID starting with 7838947abe66ed56be8aa8a24568bc6664ac0e9e78db77b03a4832e47abd3c05 not found: ID does not exist" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.791030 4923 scope.go:117] "RemoveContainer" containerID="04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.805891 4923 scope.go:117] "RemoveContainer" containerID="7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.822992 4923 scope.go:117] "RemoveContainer" containerID="001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.838230 4923 scope.go:117] "RemoveContainer" containerID="04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042" Oct 09 10:09:09 crc kubenswrapper[4923]: E1009 10:09:09.838722 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042\": container with ID starting with 04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042 not found: ID does not exist" containerID="04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.838773 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042"} err="failed to get container status \"04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042\": rpc error: code = NotFound desc = could not find container \"04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042\": container with ID starting with 04db822180d3a805eb03c16ef606e50e32ca4ad1e510541cddda51c00528d042 not found: ID does not exist" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.838806 4923 scope.go:117] "RemoveContainer" containerID="7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8" Oct 09 10:09:09 crc kubenswrapper[4923]: E1009 10:09:09.839161 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8\": container with ID starting with 7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8 not found: ID does not exist" containerID="7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.839206 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8"} err="failed to get container status \"7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8\": rpc error: code = NotFound desc = could not find container \"7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8\": container with ID starting with 7e5fc8f3396d4b7728d9a84758b31b635224f05afa6aa29704039527a30d15d8 not found: ID does not exist" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.839237 4923 scope.go:117] "RemoveContainer" containerID="001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d" Oct 09 10:09:09 crc kubenswrapper[4923]: E1009 10:09:09.839540 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d\": container with ID starting with 001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d not found: ID does not exist" containerID="001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d" Oct 09 10:09:09 crc kubenswrapper[4923]: I1009 10:09:09.839565 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d"} err="failed to get container status \"001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d\": rpc error: code = NotFound desc = could not find container \"001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d\": container with ID starting with 001944897e3a283a1f413d84e8ee1393c1712c4475f5e8b411f495041edb525d not found: ID does not exist" Oct 09 10:09:10 crc kubenswrapper[4923]: I1009 10:09:10.608336 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" path="/var/lib/kubelet/pods/31d5c282-dd47-420c-bc6e-cfcbee938bfc/volumes" Oct 09 10:09:10 crc kubenswrapper[4923]: I1009 10:09:10.609031 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" path="/var/lib/kubelet/pods/e46ac356-170c-4c91-b8b6-36ec0dfb93d8/volumes" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.041426 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkxzd"] Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.041835 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zkxzd" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerName="registry-server" containerID="cri-o://120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91" gracePeriod=2 Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.451931 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.570826 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-utilities\") pod \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.570915 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-catalog-content\") pod \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.571048 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4pj8\" (UniqueName: \"kubernetes.io/projected/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-kube-api-access-p4pj8\") pod \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\" (UID: \"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95\") " Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.571823 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-utilities" (OuterVolumeSpecName: "utilities") pod "8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" (UID: "8c0567db-2f9d-46cf-bfc8-b3893c2a4c95"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.574998 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-kube-api-access-p4pj8" (OuterVolumeSpecName: "kube-api-access-p4pj8") pod "8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" (UID: "8c0567db-2f9d-46cf-bfc8-b3893c2a4c95"). InnerVolumeSpecName "kube-api-access-p4pj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.582770 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" (UID: "8c0567db-2f9d-46cf-bfc8-b3893c2a4c95"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.672908 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4pj8\" (UniqueName: \"kubernetes.io/projected/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-kube-api-access-p4pj8\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.672949 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.672967 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.697273 4923 generic.go:334] "Generic (PLEG): container finished" podID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerID="120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91" exitCode=0 Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.697309 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkxzd" event={"ID":"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95","Type":"ContainerDied","Data":"120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91"} Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.697358 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkxzd" event={"ID":"8c0567db-2f9d-46cf-bfc8-b3893c2a4c95","Type":"ContainerDied","Data":"b410f7087b53f5548f197dc6bfb18cb6ba64e0ecd6aba5380e9da42b2003cb94"} Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.697378 4923 scope.go:117] "RemoveContainer" containerID="120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.697957 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkxzd" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.712446 4923 scope.go:117] "RemoveContainer" containerID="de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.726571 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkxzd"] Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.727883 4923 scope.go:117] "RemoveContainer" containerID="733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.730352 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkxzd"] Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.764314 4923 scope.go:117] "RemoveContainer" containerID="120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91" Oct 09 10:09:11 crc kubenswrapper[4923]: E1009 10:09:11.764804 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91\": container with ID starting with 120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91 not found: ID does not exist" containerID="120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.764947 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91"} err="failed to get container status \"120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91\": rpc error: code = NotFound desc = could not find container \"120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91\": container with ID starting with 120f0de0c0185665d2e85da8693e9f813c9dbc49a32523aa0611b98d4a444b91 not found: ID does not exist" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.765024 4923 scope.go:117] "RemoveContainer" containerID="de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1" Oct 09 10:09:11 crc kubenswrapper[4923]: E1009 10:09:11.765311 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1\": container with ID starting with de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1 not found: ID does not exist" containerID="de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.765337 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1"} err="failed to get container status \"de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1\": rpc error: code = NotFound desc = could not find container \"de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1\": container with ID starting with de5bc331a09ae73b8b07649d5982aff2d6882d94d1b75d62d0f3ecfae720fcd1 not found: ID does not exist" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.765352 4923 scope.go:117] "RemoveContainer" containerID="733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb" Oct 09 10:09:11 crc kubenswrapper[4923]: E1009 10:09:11.765804 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb\": container with ID starting with 733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb not found: ID does not exist" containerID="733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb" Oct 09 10:09:11 crc kubenswrapper[4923]: I1009 10:09:11.765834 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb"} err="failed to get container status \"733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb\": rpc error: code = NotFound desc = could not find container \"733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb\": container with ID starting with 733dab62610984e03b5320e0306a1180d57a52f8186649290621ef5cc58816bb not found: ID does not exist" Oct 09 10:09:12 crc kubenswrapper[4923]: I1009 10:09:12.609305 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" path="/var/lib/kubelet/pods/8c0567db-2f9d-46cf-bfc8-b3893c2a4c95/volumes" Oct 09 10:09:27 crc kubenswrapper[4923]: I1009 10:09:27.065934 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" podUID="eb1393c2-0fc7-4150-a96e-d48ec1d45992" containerName="oauth-openshift" containerID="cri-o://a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25" gracePeriod=15 Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.592162 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.631067 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-666545c866-8bbwc"] Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.631781 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.631794 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.631804 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerName="extract-content" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.631811 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerName="extract-content" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.631821 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.631827 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.631837 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerName="extract-utilities" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.633859 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerName="extract-utilities" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.633929 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" containerName="extract-content" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.633942 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" containerName="extract-content" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.633952 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerName="extract-content" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.633964 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerName="extract-content" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.633988 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a72feac0-410a-4cef-9d6c-f70bd8988482" containerName="pruner" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.633995 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="a72feac0-410a-4cef-9d6c-f70bd8988482" containerName="pruner" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.634008 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb1393c2-0fc7-4150-a96e-d48ec1d45992" containerName="oauth-openshift" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634014 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb1393c2-0fc7-4150-a96e-d48ec1d45992" containerName="oauth-openshift" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.634025 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634032 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.634041 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerName="extract-content" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634047 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerName="extract-content" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.634066 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerName="extract-utilities" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634077 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerName="extract-utilities" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.634086 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" containerName="extract-utilities" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634093 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" containerName="extract-utilities" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.634102 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef2f85bb-dede-4bd6-993f-cf7e80a9227f" containerName="pruner" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634109 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef2f85bb-dede-4bd6-993f-cf7e80a9227f" containerName="pruner" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.634120 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634128 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.634138 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerName="extract-utilities" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634145 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerName="extract-utilities" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634410 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="a72feac0-410a-4cef-9d6c-f70bd8988482" containerName="pruner" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634425 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c0567db-2f9d-46cf-bfc8-b3893c2a4c95" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634437 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef2f85bb-dede-4bd6-993f-cf7e80a9227f" containerName="pruner" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634444 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb1393c2-0fc7-4150-a96e-d48ec1d45992" containerName="oauth-openshift" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634454 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e46ac356-170c-4c91-b8b6-36ec0dfb93d8" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634467 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="31d5c282-dd47-420c-bc6e-cfcbee938bfc" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.634477 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="66135764-cd30-4365-bc33-c1fb96eedc85" containerName="registry-server" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.635082 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.653863 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-666545c866-8bbwc"] Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.713596 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-session\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.713915 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-provider-selection\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.714045 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-dir\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.714144 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.714183 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-ocp-branding-template\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.714377 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-router-certs\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.714486 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pfkn8\" (UniqueName: \"kubernetes.io/projected/eb1393c2-0fc7-4150-a96e-d48ec1d45992-kube-api-access-pfkn8\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.714597 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-trusted-ca-bundle\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.714719 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-serving-cert\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.714846 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-idp-0-file-data\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.714949 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-error\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.715584 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-policies\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.715723 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-service-ca\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.715842 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.715889 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-login\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.716211 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-cliconfig\") pod \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\" (UID: \"eb1393c2-0fc7-4150-a96e-d48ec1d45992\") " Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.716220 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.716483 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.716556 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-template-login\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.716654 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-template-error\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.716734 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jghpz\" (UniqueName: \"kubernetes.io/projected/87e41c13-ef2d-465c-ad1c-c3eb06b07602-kube-api-access-jghpz\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717008 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-audit-policies\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717074 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-service-ca\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717137 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/87e41c13-ef2d-465c-ad1c-c3eb06b07602-audit-dir\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717193 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717238 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-cliconfig\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717268 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717342 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717391 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717531 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717734 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-router-certs\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717843 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-session\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.717971 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-serving-cert\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.718166 4923 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.718207 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.718233 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.718260 4923 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eb1393c2-0fc7-4150-a96e-d48ec1d45992-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.718288 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.721530 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.722554 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.725999 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.726371 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.727039 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.727063 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.727191 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb1393c2-0fc7-4150-a96e-d48ec1d45992-kube-api-access-pfkn8" (OuterVolumeSpecName: "kube-api-access-pfkn8") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "kube-api-access-pfkn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.728096 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.728392 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "eb1393c2-0fc7-4150-a96e-d48ec1d45992" (UID: "eb1393c2-0fc7-4150-a96e-d48ec1d45992"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.808242 4923 generic.go:334] "Generic (PLEG): container finished" podID="eb1393c2-0fc7-4150-a96e-d48ec1d45992" containerID="a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25" exitCode=0 Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.808286 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" event={"ID":"eb1393c2-0fc7-4150-a96e-d48ec1d45992","Type":"ContainerDied","Data":"a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25"} Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.808427 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" event={"ID":"eb1393c2-0fc7-4150-a96e-d48ec1d45992","Type":"ContainerDied","Data":"42dbe3de33ad3f759f8c3a69c0d12b0108e922bd6777cec7ee7d1825b08e7cdc"} Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.808380 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-lwzgh" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.808448 4923 scope.go:117] "RemoveContainer" containerID="a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.819834 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/87e41c13-ef2d-465c-ad1c-c3eb06b07602-audit-dir\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.819922 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.819970 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-cliconfig\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820002 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820049 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820092 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820154 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-router-certs\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820198 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-session\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820237 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-serving-cert\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820282 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-template-login\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820325 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-template-error\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820366 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jghpz\" (UniqueName: \"kubernetes.io/projected/87e41c13-ef2d-465c-ad1c-c3eb06b07602-kube-api-access-jghpz\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820399 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-audit-policies\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820440 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-service-ca\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820512 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pfkn8\" (UniqueName: \"kubernetes.io/projected/eb1393c2-0fc7-4150-a96e-d48ec1d45992-kube-api-access-pfkn8\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820534 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820555 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820573 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820592 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820611 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820633 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820653 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.820672 4923 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/eb1393c2-0fc7-4150-a96e-d48ec1d45992-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.821007 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-cliconfig\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.821101 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/87e41c13-ef2d-465c-ad1c-c3eb06b07602-audit-dir\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.821745 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-service-ca\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.822736 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.822777 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/87e41c13-ef2d-465c-ad1c-c3eb06b07602-audit-policies\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.826117 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.827410 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-template-error\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.827540 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-serving-cert\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.827979 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.828485 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-template-login\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.829375 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-session\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.831299 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-system-router-certs\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.838648 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/87e41c13-ef2d-465c-ad1c-c3eb06b07602-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.842623 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jghpz\" (UniqueName: \"kubernetes.io/projected/87e41c13-ef2d-465c-ad1c-c3eb06b07602-kube-api-access-jghpz\") pod \"oauth-openshift-666545c866-8bbwc\" (UID: \"87e41c13-ef2d-465c-ad1c-c3eb06b07602\") " pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.871328 4923 scope.go:117] "RemoveContainer" containerID="a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25" Oct 09 10:09:28 crc kubenswrapper[4923]: E1009 10:09:28.871819 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25\": container with ID starting with a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25 not found: ID does not exist" containerID="a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.871854 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25"} err="failed to get container status \"a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25\": rpc error: code = NotFound desc = could not find container \"a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25\": container with ID starting with a2ff6e67f267c73a968b8fdb8ffab9006bd964622297c5f2ee49c8bb84f44a25 not found: ID does not exist" Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.888030 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lwzgh"] Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.892584 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-lwzgh"] Oct 09 10:09:28 crc kubenswrapper[4923]: I1009 10:09:28.968946 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:29 crc kubenswrapper[4923]: I1009 10:09:29.178039 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-666545c866-8bbwc"] Oct 09 10:09:29 crc kubenswrapper[4923]: I1009 10:09:29.817843 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" event={"ID":"87e41c13-ef2d-465c-ad1c-c3eb06b07602","Type":"ContainerStarted","Data":"301bc5ce041f612dc3c7443b8d12702a17b1e7534a18793fd7a4f9729594dd6a"} Oct 09 10:09:29 crc kubenswrapper[4923]: I1009 10:09:29.819870 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:29 crc kubenswrapper[4923]: I1009 10:09:29.819945 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" event={"ID":"87e41c13-ef2d-465c-ad1c-c3eb06b07602","Type":"ContainerStarted","Data":"08e92b4cdbcdf6d6d49e72929b0814c115c040ea66d05df1d7c024b8162ef5c4"} Oct 09 10:09:29 crc kubenswrapper[4923]: I1009 10:09:29.843286 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" podStartSLOduration=27.843259992 podStartE2EDuration="27.843259992s" podCreationTimestamp="2025-10-09 10:09:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:09:29.841919815 +0000 UTC m=+255.910101571" watchObservedRunningTime="2025-10-09 10:09:29.843259992 +0000 UTC m=+255.911441738" Oct 09 10:09:30 crc kubenswrapper[4923]: I1009 10:09:30.486860 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-666545c866-8bbwc" Oct 09 10:09:30 crc kubenswrapper[4923]: I1009 10:09:30.609613 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb1393c2-0fc7-4150-a96e-d48ec1d45992" path="/var/lib/kubelet/pods/eb1393c2-0fc7-4150-a96e-d48ec1d45992/volumes" Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.940587 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r7pbq"] Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.941533 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r7pbq" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerName="registry-server" containerID="cri-o://4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9" gracePeriod=30 Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.946405 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b8fd8"] Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.946661 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b8fd8" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerName="registry-server" containerID="cri-o://427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51" gracePeriod=30 Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.955020 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-897kb"] Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.955242 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" podUID="9d11e36b-e397-47e6-a3d1-93cf9367f83f" containerName="marketplace-operator" containerID="cri-o://7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719" gracePeriod=30 Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.964836 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vs8dt"] Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.965100 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vs8dt" podUID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerName="registry-server" containerID="cri-o://bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1" gracePeriod=30 Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.976818 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k9clm"] Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.977624 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.984162 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nmsxl"] Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.984395 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nmsxl" podUID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerName="registry-server" containerID="cri-o://c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533" gracePeriod=30 Oct 09 10:09:49 crc kubenswrapper[4923]: I1009 10:09:49.992884 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k9clm"] Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.105168 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a5511d1b-1e03-4020-823d-1259256d7123-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k9clm\" (UID: \"a5511d1b-1e03-4020-823d-1259256d7123\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.105281 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a5511d1b-1e03-4020-823d-1259256d7123-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k9clm\" (UID: \"a5511d1b-1e03-4020-823d-1259256d7123\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.105351 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kx62\" (UniqueName: \"kubernetes.io/projected/a5511d1b-1e03-4020-823d-1259256d7123-kube-api-access-6kx62\") pod \"marketplace-operator-79b997595-k9clm\" (UID: \"a5511d1b-1e03-4020-823d-1259256d7123\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.206247 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a5511d1b-1e03-4020-823d-1259256d7123-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k9clm\" (UID: \"a5511d1b-1e03-4020-823d-1259256d7123\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.206299 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a5511d1b-1e03-4020-823d-1259256d7123-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k9clm\" (UID: \"a5511d1b-1e03-4020-823d-1259256d7123\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.206326 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kx62\" (UniqueName: \"kubernetes.io/projected/a5511d1b-1e03-4020-823d-1259256d7123-kube-api-access-6kx62\") pod \"marketplace-operator-79b997595-k9clm\" (UID: \"a5511d1b-1e03-4020-823d-1259256d7123\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.207916 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a5511d1b-1e03-4020-823d-1259256d7123-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-k9clm\" (UID: \"a5511d1b-1e03-4020-823d-1259256d7123\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.212697 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a5511d1b-1e03-4020-823d-1259256d7123-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-k9clm\" (UID: \"a5511d1b-1e03-4020-823d-1259256d7123\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.224177 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kx62\" (UniqueName: \"kubernetes.io/projected/a5511d1b-1e03-4020-823d-1259256d7123-kube-api-access-6kx62\") pod \"marketplace-operator-79b997595-k9clm\" (UID: \"a5511d1b-1e03-4020-823d-1259256d7123\") " pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.295826 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.448058 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.480232 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.492274 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.499490 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.511400 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-operator-metrics\") pod \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.511515 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72rqw\" (UniqueName: \"kubernetes.io/projected/9d11e36b-e397-47e6-a3d1-93cf9367f83f-kube-api-access-72rqw\") pod \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.511614 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-trusted-ca\") pod \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\" (UID: \"9d11e36b-e397-47e6-a3d1-93cf9367f83f\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.512462 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "9d11e36b-e397-47e6-a3d1-93cf9367f83f" (UID: "9d11e36b-e397-47e6-a3d1-93cf9367f83f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.518606 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d11e36b-e397-47e6-a3d1-93cf9367f83f-kube-api-access-72rqw" (OuterVolumeSpecName: "kube-api-access-72rqw") pod "9d11e36b-e397-47e6-a3d1-93cf9367f83f" (UID: "9d11e36b-e397-47e6-a3d1-93cf9367f83f"). InnerVolumeSpecName "kube-api-access-72rqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.519481 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "9d11e36b-e397-47e6-a3d1-93cf9367f83f" (UID: "9d11e36b-e397-47e6-a3d1-93cf9367f83f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.590226 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.614001 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl8tr\" (UniqueName: \"kubernetes.io/projected/d4a3f04f-d392-4065-896b-7db7683f50e6-kube-api-access-kl8tr\") pod \"d4a3f04f-d392-4065-896b-7db7683f50e6\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.614074 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-catalog-content\") pod \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.614116 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-utilities\") pod \"39ceb27f-3ad7-452a-bbdd-92057099d011\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.614244 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-catalog-content\") pod \"39ceb27f-3ad7-452a-bbdd-92057099d011\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.614279 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-catalog-content\") pod \"d4a3f04f-d392-4065-896b-7db7683f50e6\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.614341 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmhct\" (UniqueName: \"kubernetes.io/projected/39ceb27f-3ad7-452a-bbdd-92057099d011-kube-api-access-qmhct\") pod \"39ceb27f-3ad7-452a-bbdd-92057099d011\" (UID: \"39ceb27f-3ad7-452a-bbdd-92057099d011\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.614408 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-utilities\") pod \"d4a3f04f-d392-4065-896b-7db7683f50e6\" (UID: \"d4a3f04f-d392-4065-896b-7db7683f50e6\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.614517 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f25d5\" (UniqueName: \"kubernetes.io/projected/09e8da7b-adaf-49f9-9811-9ca90e2bff56-kube-api-access-f25d5\") pod \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.614569 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-utilities\") pod \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\" (UID: \"09e8da7b-adaf-49f9-9811-9ca90e2bff56\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.615546 4923 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.615646 4923 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/9d11e36b-e397-47e6-a3d1-93cf9367f83f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.615692 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72rqw\" (UniqueName: \"kubernetes.io/projected/9d11e36b-e397-47e6-a3d1-93cf9367f83f-kube-api-access-72rqw\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.618652 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-utilities" (OuterVolumeSpecName: "utilities") pod "09e8da7b-adaf-49f9-9811-9ca90e2bff56" (UID: "09e8da7b-adaf-49f9-9811-9ca90e2bff56"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.621897 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-utilities" (OuterVolumeSpecName: "utilities") pod "39ceb27f-3ad7-452a-bbdd-92057099d011" (UID: "39ceb27f-3ad7-452a-bbdd-92057099d011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.622519 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-utilities" (OuterVolumeSpecName: "utilities") pod "d4a3f04f-d392-4065-896b-7db7683f50e6" (UID: "d4a3f04f-d392-4065-896b-7db7683f50e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.626203 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09e8da7b-adaf-49f9-9811-9ca90e2bff56-kube-api-access-f25d5" (OuterVolumeSpecName: "kube-api-access-f25d5") pod "09e8da7b-adaf-49f9-9811-9ca90e2bff56" (UID: "09e8da7b-adaf-49f9-9811-9ca90e2bff56"). InnerVolumeSpecName "kube-api-access-f25d5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.628680 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39ceb27f-3ad7-452a-bbdd-92057099d011-kube-api-access-qmhct" (OuterVolumeSpecName: "kube-api-access-qmhct") pod "39ceb27f-3ad7-452a-bbdd-92057099d011" (UID: "39ceb27f-3ad7-452a-bbdd-92057099d011"). InnerVolumeSpecName "kube-api-access-qmhct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.636145 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a3f04f-d392-4065-896b-7db7683f50e6-kube-api-access-kl8tr" (OuterVolumeSpecName: "kube-api-access-kl8tr") pod "d4a3f04f-d392-4065-896b-7db7683f50e6" (UID: "d4a3f04f-d392-4065-896b-7db7683f50e6"). InnerVolumeSpecName "kube-api-access-kl8tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.653329 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09e8da7b-adaf-49f9-9811-9ca90e2bff56" (UID: "09e8da7b-adaf-49f9-9811-9ca90e2bff56"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.668333 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-k9clm"] Oct 09 10:09:50 crc kubenswrapper[4923]: W1009 10:09:50.681900 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5511d1b_1e03_4020_823d_1259256d7123.slice/crio-e7c0947c9a84f0f3bfe5ccb06b96759e96660d66eaa215135c5111f351dd3ed6 WatchSource:0}: Error finding container e7c0947c9a84f0f3bfe5ccb06b96759e96660d66eaa215135c5111f351dd3ed6: Status 404 returned error can't find the container with id e7c0947c9a84f0f3bfe5ccb06b96759e96660d66eaa215135c5111f351dd3ed6 Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.686224 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39ceb27f-3ad7-452a-bbdd-92057099d011" (UID: "39ceb27f-3ad7-452a-bbdd-92057099d011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.699580 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d4a3f04f-d392-4065-896b-7db7683f50e6" (UID: "d4a3f04f-d392-4065-896b-7db7683f50e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.716585 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-utilities\") pod \"fa810245-ceec-4d06-b02d-20e39d65ff1f\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.716832 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xksvv\" (UniqueName: \"kubernetes.io/projected/fa810245-ceec-4d06-b02d-20e39d65ff1f-kube-api-access-xksvv\") pod \"fa810245-ceec-4d06-b02d-20e39d65ff1f\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.716858 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-catalog-content\") pod \"fa810245-ceec-4d06-b02d-20e39d65ff1f\" (UID: \"fa810245-ceec-4d06-b02d-20e39d65ff1f\") " Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717059 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717079 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717091 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmhct\" (UniqueName: \"kubernetes.io/projected/39ceb27f-3ad7-452a-bbdd-92057099d011-kube-api-access-qmhct\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717103 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d4a3f04f-d392-4065-896b-7db7683f50e6-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717114 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f25d5\" (UniqueName: \"kubernetes.io/projected/09e8da7b-adaf-49f9-9811-9ca90e2bff56-kube-api-access-f25d5\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717123 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717132 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kl8tr\" (UniqueName: \"kubernetes.io/projected/d4a3f04f-d392-4065-896b-7db7683f50e6-kube-api-access-kl8tr\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717139 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09e8da7b-adaf-49f9-9811-9ca90e2bff56-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717193 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39ceb27f-3ad7-452a-bbdd-92057099d011-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.717453 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-utilities" (OuterVolumeSpecName: "utilities") pod "fa810245-ceec-4d06-b02d-20e39d65ff1f" (UID: "fa810245-ceec-4d06-b02d-20e39d65ff1f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.721975 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa810245-ceec-4d06-b02d-20e39d65ff1f-kube-api-access-xksvv" (OuterVolumeSpecName: "kube-api-access-xksvv") pod "fa810245-ceec-4d06-b02d-20e39d65ff1f" (UID: "fa810245-ceec-4d06-b02d-20e39d65ff1f"). InnerVolumeSpecName "kube-api-access-xksvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.810902 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fa810245-ceec-4d06-b02d-20e39d65ff1f" (UID: "fa810245-ceec-4d06-b02d-20e39d65ff1f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.818361 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.818387 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fa810245-ceec-4d06-b02d-20e39d65ff1f-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.818402 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xksvv\" (UniqueName: \"kubernetes.io/projected/fa810245-ceec-4d06-b02d-20e39d65ff1f-kube-api-access-xksvv\") on node \"crc\" DevicePath \"\"" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.934238 4923 generic.go:334] "Generic (PLEG): container finished" podID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerID="4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9" exitCode=0 Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.934308 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7pbq" event={"ID":"39ceb27f-3ad7-452a-bbdd-92057099d011","Type":"ContainerDied","Data":"4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.934339 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r7pbq" event={"ID":"39ceb27f-3ad7-452a-bbdd-92057099d011","Type":"ContainerDied","Data":"859f2bd01c90c5a8a4c36e115e64b6e118b66065bd007cc9ce18f70bbe76218e"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.934361 4923 scope.go:117] "RemoveContainer" containerID="4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.934494 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r7pbq" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.941544 4923 generic.go:334] "Generic (PLEG): container finished" podID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerID="c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533" exitCode=0 Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.941623 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmsxl" event={"ID":"fa810245-ceec-4d06-b02d-20e39d65ff1f","Type":"ContainerDied","Data":"c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.941654 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nmsxl" event={"ID":"fa810245-ceec-4d06-b02d-20e39d65ff1f","Type":"ContainerDied","Data":"8d47c8fcd18bc8bf033991c89b78cb3c68ab0b886100b0714be53445199b5583"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.941735 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nmsxl" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.946845 4923 generic.go:334] "Generic (PLEG): container finished" podID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerID="427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51" exitCode=0 Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.946925 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8fd8" event={"ID":"d4a3f04f-d392-4065-896b-7db7683f50e6","Type":"ContainerDied","Data":"427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.946954 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b8fd8" event={"ID":"d4a3f04f-d392-4065-896b-7db7683f50e6","Type":"ContainerDied","Data":"64dc7a263529a16a4fed7c0875b966a524a42a262bc8308b578982fab80df067"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.947056 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b8fd8" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.950844 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" event={"ID":"a5511d1b-1e03-4020-823d-1259256d7123","Type":"ContainerStarted","Data":"8f6836774f3efa6cb5ff082e2036a5aed4de27fb1772105a74a53d74c351dc84"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.950901 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" event={"ID":"a5511d1b-1e03-4020-823d-1259256d7123","Type":"ContainerStarted","Data":"e7c0947c9a84f0f3bfe5ccb06b96759e96660d66eaa215135c5111f351dd3ed6"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.951735 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.952782 4923 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-k9clm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" start-of-body= Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.952817 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" podUID="a5511d1b-1e03-4020-823d-1259256d7123" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.55:8080/healthz\": dial tcp 10.217.0.55:8080: connect: connection refused" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.953058 4923 scope.go:117] "RemoveContainer" containerID="50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.956297 4923 generic.go:334] "Generic (PLEG): container finished" podID="9d11e36b-e397-47e6-a3d1-93cf9367f83f" containerID="7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719" exitCode=0 Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.956418 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" event={"ID":"9d11e36b-e397-47e6-a3d1-93cf9367f83f","Type":"ContainerDied","Data":"7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.956454 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" event={"ID":"9d11e36b-e397-47e6-a3d1-93cf9367f83f","Type":"ContainerDied","Data":"39816ad0baa2267a1a0eb13eb5ce7a75744746c39da2fb86aaccda1057b610b7"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.957237 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-897kb" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.966011 4923 generic.go:334] "Generic (PLEG): container finished" podID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerID="bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1" exitCode=0 Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.966117 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vs8dt" event={"ID":"09e8da7b-adaf-49f9-9811-9ca90e2bff56","Type":"ContainerDied","Data":"bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.966152 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vs8dt" event={"ID":"09e8da7b-adaf-49f9-9811-9ca90e2bff56","Type":"ContainerDied","Data":"277cc13a36b5fcf09a91e275a89fb9fb167e1206b4e2d28f8354b0611c351af2"} Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.966281 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vs8dt" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.975923 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" podStartSLOduration=1.97588283 podStartE2EDuration="1.97588283s" podCreationTimestamp="2025-10-09 10:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:09:50.969908199 +0000 UTC m=+277.038089965" watchObservedRunningTime="2025-10-09 10:09:50.97588283 +0000 UTC m=+277.044064586" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.987544 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r7pbq"] Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.987740 4923 scope.go:117] "RemoveContainer" containerID="ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e" Oct 09 10:09:50 crc kubenswrapper[4923]: I1009 10:09:50.995381 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r7pbq"] Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.001191 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-897kb"] Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.007827 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-897kb"] Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.024494 4923 scope.go:117] "RemoveContainer" containerID="4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.024614 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b8fd8"] Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.024675 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b8fd8"] Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.026023 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9\": container with ID starting with 4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9 not found: ID does not exist" containerID="4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.026083 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9"} err="failed to get container status \"4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9\": rpc error: code = NotFound desc = could not find container \"4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9\": container with ID starting with 4077add5303d9620c5c30e0b45ae23fe85e243c280c7a3865d24f3913b8f33b9 not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.026113 4923 scope.go:117] "RemoveContainer" containerID="50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.027206 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b\": container with ID starting with 50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b not found: ID does not exist" containerID="50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.027238 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b"} err="failed to get container status \"50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b\": rpc error: code = NotFound desc = could not find container \"50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b\": container with ID starting with 50ed2298740e12b8d80575207429cf429f29a481392ff9796caebfbdb8332d5b not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.027264 4923 scope.go:117] "RemoveContainer" containerID="ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.029830 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e\": container with ID starting with ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e not found: ID does not exist" containerID="ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.029861 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e"} err="failed to get container status \"ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e\": rpc error: code = NotFound desc = could not find container \"ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e\": container with ID starting with ae7b7ab52e2442438110032224f95b6072329e3210fa05cf995104a37dc8e33e not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.029894 4923 scope.go:117] "RemoveContainer" containerID="c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.031745 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vs8dt"] Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.035616 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vs8dt"] Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.045860 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nmsxl"] Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.050803 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nmsxl"] Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.051609 4923 scope.go:117] "RemoveContainer" containerID="a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.066651 4923 scope.go:117] "RemoveContainer" containerID="074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.078367 4923 scope.go:117] "RemoveContainer" containerID="c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.078656 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533\": container with ID starting with c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533 not found: ID does not exist" containerID="c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.078692 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533"} err="failed to get container status \"c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533\": rpc error: code = NotFound desc = could not find container \"c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533\": container with ID starting with c803e15a096f2d079811b4daeaa6961d8182e522f8d89809a2fc76f5c265e533 not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.078717 4923 scope.go:117] "RemoveContainer" containerID="a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.079103 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10\": container with ID starting with a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10 not found: ID does not exist" containerID="a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.079126 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10"} err="failed to get container status \"a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10\": rpc error: code = NotFound desc = could not find container \"a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10\": container with ID starting with a9faae087588f69606e5c9d0d4364511cea2f45abfdad3acecdfaaa1b0790b10 not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.079138 4923 scope.go:117] "RemoveContainer" containerID="074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.079336 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35\": container with ID starting with 074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35 not found: ID does not exist" containerID="074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.079360 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35"} err="failed to get container status \"074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35\": rpc error: code = NotFound desc = could not find container \"074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35\": container with ID starting with 074ea917122df9cc5edd00062ce8d9e47f7108ef42c97e9d3bcfde3706a99b35 not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.079378 4923 scope.go:117] "RemoveContainer" containerID="427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.138342 4923 scope.go:117] "RemoveContainer" containerID="a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.155785 4923 scope.go:117] "RemoveContainer" containerID="8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.169499 4923 scope.go:117] "RemoveContainer" containerID="427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.170287 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51\": container with ID starting with 427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51 not found: ID does not exist" containerID="427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.170316 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51"} err="failed to get container status \"427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51\": rpc error: code = NotFound desc = could not find container \"427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51\": container with ID starting with 427dde5af91a99c6978bf6df47df0a12f37a9e8ce57981bb17653bf653aefc51 not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.170339 4923 scope.go:117] "RemoveContainer" containerID="a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.171516 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e\": container with ID starting with a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e not found: ID does not exist" containerID="a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.171553 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e"} err="failed to get container status \"a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e\": rpc error: code = NotFound desc = could not find container \"a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e\": container with ID starting with a8d474d46557a0d3429b7b8ac4ef56d967e3fa6f3cab7ba44f74908631a7495e not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.171573 4923 scope.go:117] "RemoveContainer" containerID="8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.172125 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2\": container with ID starting with 8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2 not found: ID does not exist" containerID="8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.172149 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2"} err="failed to get container status \"8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2\": rpc error: code = NotFound desc = could not find container \"8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2\": container with ID starting with 8f79014c7f3d35dabd1f95d81dd150c3c10ada3003cbc850662aefed05948fa2 not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.172167 4923 scope.go:117] "RemoveContainer" containerID="7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.186614 4923 scope.go:117] "RemoveContainer" containerID="7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.187670 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719\": container with ID starting with 7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719 not found: ID does not exist" containerID="7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.187700 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719"} err="failed to get container status \"7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719\": rpc error: code = NotFound desc = could not find container \"7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719\": container with ID starting with 7b0a9fb9052d92b98691c90034b2b37c8cc51e032e4447abc4622fbec954f719 not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.187721 4923 scope.go:117] "RemoveContainer" containerID="bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.201725 4923 scope.go:117] "RemoveContainer" containerID="2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.216604 4923 scope.go:117] "RemoveContainer" containerID="83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.231319 4923 scope.go:117] "RemoveContainer" containerID="bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.231903 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1\": container with ID starting with bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1 not found: ID does not exist" containerID="bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.231959 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1"} err="failed to get container status \"bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1\": rpc error: code = NotFound desc = could not find container \"bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1\": container with ID starting with bef2cb2fa5b42220d2d83d6d05d94b9e16590f64fd43445dbc5505b6377084f1 not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.231992 4923 scope.go:117] "RemoveContainer" containerID="2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.232294 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f\": container with ID starting with 2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f not found: ID does not exist" containerID="2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.232320 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f"} err="failed to get container status \"2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f\": rpc error: code = NotFound desc = could not find container \"2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f\": container with ID starting with 2f148d5a5a3eef361cb498749c78107d7f81124c032609f36001a334f29d8e6f not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.232343 4923 scope.go:117] "RemoveContainer" containerID="83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8" Oct 09 10:09:51 crc kubenswrapper[4923]: E1009 10:09:51.232576 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8\": container with ID starting with 83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8 not found: ID does not exist" containerID="83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.232609 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8"} err="failed to get container status \"83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8\": rpc error: code = NotFound desc = could not find container \"83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8\": container with ID starting with 83eca8aa6d3452a3ff7ecb0a074929b5a5e9d6cdc4c05c7b1f4a30e9b61015c8 not found: ID does not exist" Oct 09 10:09:51 crc kubenswrapper[4923]: I1009 10:09:51.982978 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-k9clm" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143022 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hbbk5"] Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143223 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d11e36b-e397-47e6-a3d1-93cf9367f83f" containerName="marketplace-operator" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143238 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d11e36b-e397-47e6-a3d1-93cf9367f83f" containerName="marketplace-operator" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143251 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143259 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143270 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerName="extract-content" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143276 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerName="extract-content" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143284 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143290 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143298 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerName="extract-utilities" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143304 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerName="extract-utilities" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143314 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerName="extract-utilities" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143320 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerName="extract-utilities" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143327 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerName="extract-content" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143333 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerName="extract-content" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143339 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143346 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143355 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143361 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143372 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerName="extract-content" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143377 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerName="extract-content" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143389 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerName="extract-utilities" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143395 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerName="extract-utilities" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143404 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerName="extract-utilities" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143409 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerName="extract-utilities" Oct 09 10:09:52 crc kubenswrapper[4923]: E1009 10:09:52.143417 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerName="extract-content" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143422 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerName="extract-content" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143501 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143512 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143519 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa810245-ceec-4d06-b02d-20e39d65ff1f" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143526 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d11e36b-e397-47e6-a3d1-93cf9367f83f" containerName="marketplace-operator" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.143533 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" containerName="registry-server" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.144161 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.146008 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.155143 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hbbk5"] Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.236066 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-utilities\") pod \"certified-operators-hbbk5\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.236127 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-catalog-content\") pod \"certified-operators-hbbk5\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.236246 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75cdc\" (UniqueName: \"kubernetes.io/projected/09837ba0-2b96-4f41-bb93-268a05cf3f2d-kube-api-access-75cdc\") pod \"certified-operators-hbbk5\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.339082 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75cdc\" (UniqueName: \"kubernetes.io/projected/09837ba0-2b96-4f41-bb93-268a05cf3f2d-kube-api-access-75cdc\") pod \"certified-operators-hbbk5\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.339651 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-utilities\") pod \"certified-operators-hbbk5\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.339180 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-utilities\") pod \"certified-operators-hbbk5\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.339740 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-catalog-content\") pod \"certified-operators-hbbk5\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.339967 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-catalog-content\") pod \"certified-operators-hbbk5\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.344654 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-44dmh"] Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.345720 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.347905 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.359206 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-44dmh"] Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.370854 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75cdc\" (UniqueName: \"kubernetes.io/projected/09837ba0-2b96-4f41-bb93-268a05cf3f2d-kube-api-access-75cdc\") pod \"certified-operators-hbbk5\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.441843 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d773cf-afe5-4611-a180-c027b6e3c129-utilities\") pod \"redhat-marketplace-44dmh\" (UID: \"d1d773cf-afe5-4611-a180-c027b6e3c129\") " pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.442167 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d773cf-afe5-4611-a180-c027b6e3c129-catalog-content\") pod \"redhat-marketplace-44dmh\" (UID: \"d1d773cf-afe5-4611-a180-c027b6e3c129\") " pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.442469 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmhhx\" (UniqueName: \"kubernetes.io/projected/d1d773cf-afe5-4611-a180-c027b6e3c129-kube-api-access-bmhhx\") pod \"redhat-marketplace-44dmh\" (UID: \"d1d773cf-afe5-4611-a180-c027b6e3c129\") " pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.471904 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.546207 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d773cf-afe5-4611-a180-c027b6e3c129-catalog-content\") pod \"redhat-marketplace-44dmh\" (UID: \"d1d773cf-afe5-4611-a180-c027b6e3c129\") " pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.546596 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmhhx\" (UniqueName: \"kubernetes.io/projected/d1d773cf-afe5-4611-a180-c027b6e3c129-kube-api-access-bmhhx\") pod \"redhat-marketplace-44dmh\" (UID: \"d1d773cf-afe5-4611-a180-c027b6e3c129\") " pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.546639 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d773cf-afe5-4611-a180-c027b6e3c129-utilities\") pod \"redhat-marketplace-44dmh\" (UID: \"d1d773cf-afe5-4611-a180-c027b6e3c129\") " pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.547068 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d1d773cf-afe5-4611-a180-c027b6e3c129-catalog-content\") pod \"redhat-marketplace-44dmh\" (UID: \"d1d773cf-afe5-4611-a180-c027b6e3c129\") " pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.547322 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d1d773cf-afe5-4611-a180-c027b6e3c129-utilities\") pod \"redhat-marketplace-44dmh\" (UID: \"d1d773cf-afe5-4611-a180-c027b6e3c129\") " pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.571161 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmhhx\" (UniqueName: \"kubernetes.io/projected/d1d773cf-afe5-4611-a180-c027b6e3c129-kube-api-access-bmhhx\") pod \"redhat-marketplace-44dmh\" (UID: \"d1d773cf-afe5-4611-a180-c027b6e3c129\") " pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.610908 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09e8da7b-adaf-49f9-9811-9ca90e2bff56" path="/var/lib/kubelet/pods/09e8da7b-adaf-49f9-9811-9ca90e2bff56/volumes" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.611739 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39ceb27f-3ad7-452a-bbdd-92057099d011" path="/var/lib/kubelet/pods/39ceb27f-3ad7-452a-bbdd-92057099d011/volumes" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.612360 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d11e36b-e397-47e6-a3d1-93cf9367f83f" path="/var/lib/kubelet/pods/9d11e36b-e397-47e6-a3d1-93cf9367f83f/volumes" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.613234 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a3f04f-d392-4065-896b-7db7683f50e6" path="/var/lib/kubelet/pods/d4a3f04f-d392-4065-896b-7db7683f50e6/volumes" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.613803 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa810245-ceec-4d06-b02d-20e39d65ff1f" path="/var/lib/kubelet/pods/fa810245-ceec-4d06-b02d-20e39d65ff1f/volumes" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.690267 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.877060 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hbbk5"] Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.882442 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-44dmh"] Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.985392 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44dmh" event={"ID":"d1d773cf-afe5-4611-a180-c027b6e3c129","Type":"ContainerStarted","Data":"286387b01d45b58856aac6b89a90380302d9e23bb4f27b2d8ad6fa687b4184c0"} Oct 09 10:09:52 crc kubenswrapper[4923]: I1009 10:09:52.986631 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbbk5" event={"ID":"09837ba0-2b96-4f41-bb93-268a05cf3f2d","Type":"ContainerStarted","Data":"f47b11e9b70e23234c6946e6d161607f4875ccddf5c25d95161e8695c85383c4"} Oct 09 10:09:53 crc kubenswrapper[4923]: I1009 10:09:53.991583 4923 generic.go:334] "Generic (PLEG): container finished" podID="d1d773cf-afe5-4611-a180-c027b6e3c129" containerID="72bcf05288b11546e7960b3d0a3fd109ea7fc01078d173d805fec34521448967" exitCode=0 Oct 09 10:09:53 crc kubenswrapper[4923]: I1009 10:09:53.991620 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44dmh" event={"ID":"d1d773cf-afe5-4611-a180-c027b6e3c129","Type":"ContainerDied","Data":"72bcf05288b11546e7960b3d0a3fd109ea7fc01078d173d805fec34521448967"} Oct 09 10:09:53 crc kubenswrapper[4923]: I1009 10:09:53.993236 4923 generic.go:334] "Generic (PLEG): container finished" podID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerID="a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd" exitCode=0 Oct 09 10:09:53 crc kubenswrapper[4923]: I1009 10:09:53.993323 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbbk5" event={"ID":"09837ba0-2b96-4f41-bb93-268a05cf3f2d","Type":"ContainerDied","Data":"a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd"} Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.545218 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-brvm4"] Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.546545 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.548787 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.554952 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brvm4"] Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.675693 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1e82e8-6bce-4de6-a52d-348db9bf7e6e-utilities\") pod \"community-operators-brvm4\" (UID: \"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e\") " pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.675738 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxlzw\" (UniqueName: \"kubernetes.io/projected/8d1e82e8-6bce-4de6-a52d-348db9bf7e6e-kube-api-access-xxlzw\") pod \"community-operators-brvm4\" (UID: \"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e\") " pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.675953 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1e82e8-6bce-4de6-a52d-348db9bf7e6e-catalog-content\") pod \"community-operators-brvm4\" (UID: \"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e\") " pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.745313 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4btwb"] Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.746212 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.748261 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.761253 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4btwb"] Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.776811 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1e82e8-6bce-4de6-a52d-348db9bf7e6e-utilities\") pod \"community-operators-brvm4\" (UID: \"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e\") " pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.776875 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxlzw\" (UniqueName: \"kubernetes.io/projected/8d1e82e8-6bce-4de6-a52d-348db9bf7e6e-kube-api-access-xxlzw\") pod \"community-operators-brvm4\" (UID: \"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e\") " pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.776964 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1e82e8-6bce-4de6-a52d-348db9bf7e6e-catalog-content\") pod \"community-operators-brvm4\" (UID: \"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e\") " pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.777380 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d1e82e8-6bce-4de6-a52d-348db9bf7e6e-catalog-content\") pod \"community-operators-brvm4\" (UID: \"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e\") " pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.777597 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d1e82e8-6bce-4de6-a52d-348db9bf7e6e-utilities\") pod \"community-operators-brvm4\" (UID: \"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e\") " pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.800687 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxlzw\" (UniqueName: \"kubernetes.io/projected/8d1e82e8-6bce-4de6-a52d-348db9bf7e6e-kube-api-access-xxlzw\") pod \"community-operators-brvm4\" (UID: \"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e\") " pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.871317 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.877799 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ddd9f92-4734-4e5a-b5cb-7d4e7448596f-utilities\") pod \"redhat-operators-4btwb\" (UID: \"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f\") " pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.878050 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ddd9f92-4734-4e5a-b5cb-7d4e7448596f-catalog-content\") pod \"redhat-operators-4btwb\" (UID: \"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f\") " pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.878072 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wszrk\" (UniqueName: \"kubernetes.io/projected/3ddd9f92-4734-4e5a-b5cb-7d4e7448596f-kube-api-access-wszrk\") pod \"redhat-operators-4btwb\" (UID: \"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f\") " pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.978932 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ddd9f92-4734-4e5a-b5cb-7d4e7448596f-catalog-content\") pod \"redhat-operators-4btwb\" (UID: \"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f\") " pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.978984 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wszrk\" (UniqueName: \"kubernetes.io/projected/3ddd9f92-4734-4e5a-b5cb-7d4e7448596f-kube-api-access-wszrk\") pod \"redhat-operators-4btwb\" (UID: \"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f\") " pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.979077 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ddd9f92-4734-4e5a-b5cb-7d4e7448596f-utilities\") pod \"redhat-operators-4btwb\" (UID: \"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f\") " pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.980095 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3ddd9f92-4734-4e5a-b5cb-7d4e7448596f-utilities\") pod \"redhat-operators-4btwb\" (UID: \"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f\") " pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:54 crc kubenswrapper[4923]: I1009 10:09:54.980166 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3ddd9f92-4734-4e5a-b5cb-7d4e7448596f-catalog-content\") pod \"redhat-operators-4btwb\" (UID: \"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f\") " pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:55 crc kubenswrapper[4923]: I1009 10:09:55.014166 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wszrk\" (UniqueName: \"kubernetes.io/projected/3ddd9f92-4734-4e5a-b5cb-7d4e7448596f-kube-api-access-wszrk\") pod \"redhat-operators-4btwb\" (UID: \"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f\") " pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:55 crc kubenswrapper[4923]: I1009 10:09:55.067349 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:09:55 crc kubenswrapper[4923]: I1009 10:09:55.293570 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-brvm4"] Oct 09 10:09:55 crc kubenswrapper[4923]: W1009 10:09:55.298710 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8d1e82e8_6bce_4de6_a52d_348db9bf7e6e.slice/crio-638f69d573b737c24e8e067db1b7bb8e14e6db27d48c1796c07209bfd4b50933 WatchSource:0}: Error finding container 638f69d573b737c24e8e067db1b7bb8e14e6db27d48c1796c07209bfd4b50933: Status 404 returned error can't find the container with id 638f69d573b737c24e8e067db1b7bb8e14e6db27d48c1796c07209bfd4b50933 Oct 09 10:09:55 crc kubenswrapper[4923]: I1009 10:09:55.472151 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4btwb"] Oct 09 10:09:55 crc kubenswrapper[4923]: W1009 10:09:55.480677 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3ddd9f92_4734_4e5a_b5cb_7d4e7448596f.slice/crio-58292d3c3246a32828a84716d799b9e65ac2c00fd115220502931122a30ba145 WatchSource:0}: Error finding container 58292d3c3246a32828a84716d799b9e65ac2c00fd115220502931122a30ba145: Status 404 returned error can't find the container with id 58292d3c3246a32828a84716d799b9e65ac2c00fd115220502931122a30ba145 Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.005985 4923 generic.go:334] "Generic (PLEG): container finished" podID="8d1e82e8-6bce-4de6-a52d-348db9bf7e6e" containerID="be62da96c214b847e8fc1d5153ef905b80e64c39c3bea56275851a8e5e904dc6" exitCode=0 Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.006087 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvm4" event={"ID":"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e","Type":"ContainerDied","Data":"be62da96c214b847e8fc1d5153ef905b80e64c39c3bea56275851a8e5e904dc6"} Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.006318 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvm4" event={"ID":"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e","Type":"ContainerStarted","Data":"638f69d573b737c24e8e067db1b7bb8e14e6db27d48c1796c07209bfd4b50933"} Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.007979 4923 generic.go:334] "Generic (PLEG): container finished" podID="d1d773cf-afe5-4611-a180-c027b6e3c129" containerID="16c5d03c9dab7bfdd9b31c6ffc473c8373f337d20e9c84f36c9f5fc304d8006b" exitCode=0 Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.008027 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44dmh" event={"ID":"d1d773cf-afe5-4611-a180-c027b6e3c129","Type":"ContainerDied","Data":"16c5d03c9dab7bfdd9b31c6ffc473c8373f337d20e9c84f36c9f5fc304d8006b"} Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.010979 4923 generic.go:334] "Generic (PLEG): container finished" podID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerID="55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0" exitCode=0 Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.011037 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbbk5" event={"ID":"09837ba0-2b96-4f41-bb93-268a05cf3f2d","Type":"ContainerDied","Data":"55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0"} Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.014043 4923 generic.go:334] "Generic (PLEG): container finished" podID="3ddd9f92-4734-4e5a-b5cb-7d4e7448596f" containerID="52913f55301b1df41bf2c972d1739d5bf3164d8cc546a61c29de9c49d83d6323" exitCode=0 Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.014065 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4btwb" event={"ID":"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f","Type":"ContainerDied","Data":"52913f55301b1df41bf2c972d1739d5bf3164d8cc546a61c29de9c49d83d6323"} Oct 09 10:09:56 crc kubenswrapper[4923]: I1009 10:09:56.014081 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4btwb" event={"ID":"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f","Type":"ContainerStarted","Data":"58292d3c3246a32828a84716d799b9e65ac2c00fd115220502931122a30ba145"} Oct 09 10:09:57 crc kubenswrapper[4923]: I1009 10:09:57.022593 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-44dmh" event={"ID":"d1d773cf-afe5-4611-a180-c027b6e3c129","Type":"ContainerStarted","Data":"972fe095cfefeb44e0a4dcce5d1bad5df4d2a0484011536423d6ec87f3d29e82"} Oct 09 10:09:57 crc kubenswrapper[4923]: I1009 10:09:57.045267 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-44dmh" podStartSLOduration=2.258785813 podStartE2EDuration="5.045246106s" podCreationTimestamp="2025-10-09 10:09:52 +0000 UTC" firstStartedPulling="2025-10-09 10:09:53.994129828 +0000 UTC m=+280.062311574" lastFinishedPulling="2025-10-09 10:09:56.780590111 +0000 UTC m=+282.848771867" observedRunningTime="2025-10-09 10:09:57.04259027 +0000 UTC m=+283.110772046" watchObservedRunningTime="2025-10-09 10:09:57.045246106 +0000 UTC m=+283.113427862" Oct 09 10:09:58 crc kubenswrapper[4923]: I1009 10:09:58.030806 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbbk5" event={"ID":"09837ba0-2b96-4f41-bb93-268a05cf3f2d","Type":"ContainerStarted","Data":"112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4"} Oct 09 10:09:58 crc kubenswrapper[4923]: I1009 10:09:58.032628 4923 generic.go:334] "Generic (PLEG): container finished" podID="3ddd9f92-4734-4e5a-b5cb-7d4e7448596f" containerID="ab40cb76a6826a9dae6346126e1e55cae2877484c64362ead836e7a98a77469b" exitCode=0 Oct 09 10:09:58 crc kubenswrapper[4923]: I1009 10:09:58.032669 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4btwb" event={"ID":"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f","Type":"ContainerDied","Data":"ab40cb76a6826a9dae6346126e1e55cae2877484c64362ead836e7a98a77469b"} Oct 09 10:09:58 crc kubenswrapper[4923]: I1009 10:09:58.035475 4923 generic.go:334] "Generic (PLEG): container finished" podID="8d1e82e8-6bce-4de6-a52d-348db9bf7e6e" containerID="12de086440c807c9b0fada227740344916f9c9a415954707bf0aba6b8a213fba" exitCode=0 Oct 09 10:09:58 crc kubenswrapper[4923]: I1009 10:09:58.035587 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvm4" event={"ID":"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e","Type":"ContainerDied","Data":"12de086440c807c9b0fada227740344916f9c9a415954707bf0aba6b8a213fba"} Oct 09 10:09:58 crc kubenswrapper[4923]: I1009 10:09:58.048252 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hbbk5" podStartSLOduration=3.084507062 podStartE2EDuration="6.04823914s" podCreationTimestamp="2025-10-09 10:09:52 +0000 UTC" firstStartedPulling="2025-10-09 10:09:53.994958341 +0000 UTC m=+280.063140097" lastFinishedPulling="2025-10-09 10:09:56.958690419 +0000 UTC m=+283.026872175" observedRunningTime="2025-10-09 10:09:58.047090386 +0000 UTC m=+284.115272142" watchObservedRunningTime="2025-10-09 10:09:58.04823914 +0000 UTC m=+284.116420896" Oct 09 10:10:01 crc kubenswrapper[4923]: I1009 10:10:01.050929 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4btwb" event={"ID":"3ddd9f92-4734-4e5a-b5cb-7d4e7448596f","Type":"ContainerStarted","Data":"5f11883e577c313e21b579742f4e74631cf323042eb1373d98ad0931986d178c"} Oct 09 10:10:01 crc kubenswrapper[4923]: I1009 10:10:01.053331 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-brvm4" event={"ID":"8d1e82e8-6bce-4de6-a52d-348db9bf7e6e","Type":"ContainerStarted","Data":"0bae8073f3a38cb5adef6db0265d7ce1f2c90ef7e855e233f9edc1814f1b9076"} Oct 09 10:10:01 crc kubenswrapper[4923]: I1009 10:10:01.069954 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4btwb" podStartSLOduration=3.530202834 podStartE2EDuration="7.069935686s" podCreationTimestamp="2025-10-09 10:09:54 +0000 UTC" firstStartedPulling="2025-10-09 10:09:56.014811936 +0000 UTC m=+282.082993692" lastFinishedPulling="2025-10-09 10:09:59.554544788 +0000 UTC m=+285.622726544" observedRunningTime="2025-10-09 10:10:01.066414134 +0000 UTC m=+287.134595920" watchObservedRunningTime="2025-10-09 10:10:01.069935686 +0000 UTC m=+287.138117442" Oct 09 10:10:01 crc kubenswrapper[4923]: I1009 10:10:01.082574 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-brvm4" podStartSLOduration=3.213366538 podStartE2EDuration="7.082555387s" podCreationTimestamp="2025-10-09 10:09:54 +0000 UTC" firstStartedPulling="2025-10-09 10:09:56.008422764 +0000 UTC m=+282.076604530" lastFinishedPulling="2025-10-09 10:09:59.877611613 +0000 UTC m=+285.945793379" observedRunningTime="2025-10-09 10:10:01.082040612 +0000 UTC m=+287.150222368" watchObservedRunningTime="2025-10-09 10:10:01.082555387 +0000 UTC m=+287.150737143" Oct 09 10:10:02 crc kubenswrapper[4923]: I1009 10:10:02.472140 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:10:02 crc kubenswrapper[4923]: I1009 10:10:02.472467 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:10:02 crc kubenswrapper[4923]: I1009 10:10:02.519811 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:10:02 crc kubenswrapper[4923]: I1009 10:10:02.691140 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:10:02 crc kubenswrapper[4923]: I1009 10:10:02.692054 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:10:02 crc kubenswrapper[4923]: I1009 10:10:02.729365 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:10:03 crc kubenswrapper[4923]: I1009 10:10:03.101898 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-44dmh" Oct 09 10:10:03 crc kubenswrapper[4923]: I1009 10:10:03.127820 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:10:04 crc kubenswrapper[4923]: I1009 10:10:04.871825 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:10:04 crc kubenswrapper[4923]: I1009 10:10:04.872124 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:10:04 crc kubenswrapper[4923]: I1009 10:10:04.912314 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:10:05 crc kubenswrapper[4923]: I1009 10:10:05.067989 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:10:05 crc kubenswrapper[4923]: I1009 10:10:05.068103 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:10:05 crc kubenswrapper[4923]: I1009 10:10:05.104822 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:10:05 crc kubenswrapper[4923]: I1009 10:10:05.119865 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-brvm4" Oct 09 10:10:06 crc kubenswrapper[4923]: I1009 10:10:06.113679 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4btwb" Oct 09 10:11:24 crc kubenswrapper[4923]: I1009 10:11:24.599373 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:11:24 crc kubenswrapper[4923]: I1009 10:11:24.600277 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:11:54 crc kubenswrapper[4923]: I1009 10:11:54.599286 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:11:54 crc kubenswrapper[4923]: I1009 10:11:54.599845 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.150598 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-889mc"] Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.152799 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.208685 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-889mc"] Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.353699 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ee29ba8a-e56d-455c-a275-0f756ade6105-registry-tls\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.353795 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee29ba8a-e56d-455c-a275-0f756ade6105-bound-sa-token\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.353824 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ee29ba8a-e56d-455c-a275-0f756ade6105-ca-trust-extracted\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.353904 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee29ba8a-e56d-455c-a275-0f756ade6105-trusted-ca\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.353957 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ee29ba8a-e56d-455c-a275-0f756ade6105-registry-certificates\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.354021 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.354123 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwq5r\" (UniqueName: \"kubernetes.io/projected/ee29ba8a-e56d-455c-a275-0f756ade6105-kube-api-access-wwq5r\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.354203 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ee29ba8a-e56d-455c-a275-0f756ade6105-installation-pull-secrets\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.376934 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.456667 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee29ba8a-e56d-455c-a275-0f756ade6105-trusted-ca\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.456727 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ee29ba8a-e56d-455c-a275-0f756ade6105-registry-certificates\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.456793 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwq5r\" (UniqueName: \"kubernetes.io/projected/ee29ba8a-e56d-455c-a275-0f756ade6105-kube-api-access-wwq5r\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.456837 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ee29ba8a-e56d-455c-a275-0f756ade6105-installation-pull-secrets\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.456886 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ee29ba8a-e56d-455c-a275-0f756ade6105-registry-tls\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.456926 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee29ba8a-e56d-455c-a275-0f756ade6105-bound-sa-token\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.456955 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ee29ba8a-e56d-455c-a275-0f756ade6105-ca-trust-extracted\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.457512 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ee29ba8a-e56d-455c-a275-0f756ade6105-ca-trust-extracted\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.458363 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ee29ba8a-e56d-455c-a275-0f756ade6105-registry-certificates\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.458692 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ee29ba8a-e56d-455c-a275-0f756ade6105-trusted-ca\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.465148 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ee29ba8a-e56d-455c-a275-0f756ade6105-registry-tls\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.465168 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ee29ba8a-e56d-455c-a275-0f756ade6105-installation-pull-secrets\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.473398 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ee29ba8a-e56d-455c-a275-0f756ade6105-bound-sa-token\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.475213 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwq5r\" (UniqueName: \"kubernetes.io/projected/ee29ba8a-e56d-455c-a275-0f756ade6105-kube-api-access-wwq5r\") pod \"image-registry-66df7c8f76-889mc\" (UID: \"ee29ba8a-e56d-455c-a275-0f756ade6105\") " pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.771478 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:05 crc kubenswrapper[4923]: I1009 10:12:05.960236 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-889mc"] Oct 09 10:12:06 crc kubenswrapper[4923]: I1009 10:12:06.735068 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-889mc" event={"ID":"ee29ba8a-e56d-455c-a275-0f756ade6105","Type":"ContainerStarted","Data":"e929622ceae4150546e2d905ca5cc108a6409fcbcc4bfb8657b2f357f0696d5a"} Oct 09 10:12:06 crc kubenswrapper[4923]: I1009 10:12:06.737032 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-889mc" event={"ID":"ee29ba8a-e56d-455c-a275-0f756ade6105","Type":"ContainerStarted","Data":"8bb93897efcbde31836386ca9f66572cd1a7ab44f0dd9bc07ebdbbe5e72ac778"} Oct 09 10:12:06 crc kubenswrapper[4923]: I1009 10:12:06.737142 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:24 crc kubenswrapper[4923]: I1009 10:12:24.599423 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:12:24 crc kubenswrapper[4923]: I1009 10:12:24.600017 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:12:24 crc kubenswrapper[4923]: I1009 10:12:24.600102 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:12:24 crc kubenswrapper[4923]: I1009 10:12:24.600803 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fdc7d576a4e21a1ee542a4eba76ea241d8e56b0e66e0f288027bd3d2de345b25"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:12:24 crc kubenswrapper[4923]: I1009 10:12:24.600869 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://fdc7d576a4e21a1ee542a4eba76ea241d8e56b0e66e0f288027bd3d2de345b25" gracePeriod=600 Oct 09 10:12:24 crc kubenswrapper[4923]: I1009 10:12:24.838902 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="fdc7d576a4e21a1ee542a4eba76ea241d8e56b0e66e0f288027bd3d2de345b25" exitCode=0 Oct 09 10:12:24 crc kubenswrapper[4923]: I1009 10:12:24.839017 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"fdc7d576a4e21a1ee542a4eba76ea241d8e56b0e66e0f288027bd3d2de345b25"} Oct 09 10:12:24 crc kubenswrapper[4923]: I1009 10:12:24.839305 4923 scope.go:117] "RemoveContainer" containerID="dece62c6939c2e082b6b73e04df6e610bf282690040406a986e0662a1a4072f0" Oct 09 10:12:25 crc kubenswrapper[4923]: I1009 10:12:25.776862 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-889mc" Oct 09 10:12:25 crc kubenswrapper[4923]: I1009 10:12:25.798094 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-889mc" podStartSLOduration=20.798072144 podStartE2EDuration="20.798072144s" podCreationTimestamp="2025-10-09 10:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:12:06.755320218 +0000 UTC m=+412.823501974" watchObservedRunningTime="2025-10-09 10:12:25.798072144 +0000 UTC m=+431.866253910" Oct 09 10:12:25 crc kubenswrapper[4923]: I1009 10:12:25.827059 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvsh5"] Oct 09 10:12:25 crc kubenswrapper[4923]: I1009 10:12:25.846483 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"15e3211b265dc1c111dcb9faf998805ee133fefab85afa177cdecf472b4fa20c"} Oct 09 10:12:50 crc kubenswrapper[4923]: I1009 10:12:50.871320 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" podUID="02a21759-421a-429b-ac1b-c43cd7c236f0" containerName="registry" containerID="cri-o://8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407" gracePeriod=30 Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.702953 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.784026 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-trusted-ca\") pod \"02a21759-421a-429b-ac1b-c43cd7c236f0\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.784096 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/02a21759-421a-429b-ac1b-c43cd7c236f0-installation-pull-secrets\") pod \"02a21759-421a-429b-ac1b-c43cd7c236f0\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.784120 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-bound-sa-token\") pod \"02a21759-421a-429b-ac1b-c43cd7c236f0\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.784250 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"02a21759-421a-429b-ac1b-c43cd7c236f0\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.784281 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/02a21759-421a-429b-ac1b-c43cd7c236f0-ca-trust-extracted\") pod \"02a21759-421a-429b-ac1b-c43cd7c236f0\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.784302 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vh6l\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-kube-api-access-8vh6l\") pod \"02a21759-421a-429b-ac1b-c43cd7c236f0\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.784396 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-certificates\") pod \"02a21759-421a-429b-ac1b-c43cd7c236f0\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.784445 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-tls\") pod \"02a21759-421a-429b-ac1b-c43cd7c236f0\" (UID: \"02a21759-421a-429b-ac1b-c43cd7c236f0\") " Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.785164 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "02a21759-421a-429b-ac1b-c43cd7c236f0" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.786123 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "02a21759-421a-429b-ac1b-c43cd7c236f0" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.793083 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02a21759-421a-429b-ac1b-c43cd7c236f0-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "02a21759-421a-429b-ac1b-c43cd7c236f0" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.793251 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "02a21759-421a-429b-ac1b-c43cd7c236f0" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.793530 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "02a21759-421a-429b-ac1b-c43cd7c236f0" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.793922 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "02a21759-421a-429b-ac1b-c43cd7c236f0" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.795634 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-kube-api-access-8vh6l" (OuterVolumeSpecName: "kube-api-access-8vh6l") pod "02a21759-421a-429b-ac1b-c43cd7c236f0" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0"). InnerVolumeSpecName "kube-api-access-8vh6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.808866 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02a21759-421a-429b-ac1b-c43cd7c236f0-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "02a21759-421a-429b-ac1b-c43cd7c236f0" (UID: "02a21759-421a-429b-ac1b-c43cd7c236f0"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.886087 4923 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.886119 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.886130 4923 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/02a21759-421a-429b-ac1b-c43cd7c236f0-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.886142 4923 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.886152 4923 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/02a21759-421a-429b-ac1b-c43cd7c236f0-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.886160 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vh6l\" (UniqueName: \"kubernetes.io/projected/02a21759-421a-429b-ac1b-c43cd7c236f0-kube-api-access-8vh6l\") on node \"crc\" DevicePath \"\"" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.886169 4923 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/02a21759-421a-429b-ac1b-c43cd7c236f0-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.984581 4923 generic.go:334] "Generic (PLEG): container finished" podID="02a21759-421a-429b-ac1b-c43cd7c236f0" containerID="8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407" exitCode=0 Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.984645 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.984640 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" event={"ID":"02a21759-421a-429b-ac1b-c43cd7c236f0","Type":"ContainerDied","Data":"8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407"} Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.984704 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-pvsh5" event={"ID":"02a21759-421a-429b-ac1b-c43cd7c236f0","Type":"ContainerDied","Data":"6bb80ea12b6e32d97fd4826c6900da6eae9909aa253d2360d04452fc9523f4ee"} Oct 09 10:12:51 crc kubenswrapper[4923]: I1009 10:12:51.984725 4923 scope.go:117] "RemoveContainer" containerID="8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407" Oct 09 10:12:52 crc kubenswrapper[4923]: I1009 10:12:52.007137 4923 scope.go:117] "RemoveContainer" containerID="8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407" Oct 09 10:12:52 crc kubenswrapper[4923]: E1009 10:12:52.007797 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407\": container with ID starting with 8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407 not found: ID does not exist" containerID="8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407" Oct 09 10:12:52 crc kubenswrapper[4923]: I1009 10:12:52.007882 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407"} err="failed to get container status \"8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407\": rpc error: code = NotFound desc = could not find container \"8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407\": container with ID starting with 8a7323233bc7f48341f43dfe1025d3affffd624d52192e40cd392f2f42cb3407 not found: ID does not exist" Oct 09 10:12:52 crc kubenswrapper[4923]: I1009 10:12:52.016211 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvsh5"] Oct 09 10:12:52 crc kubenswrapper[4923]: I1009 10:12:52.021868 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-pvsh5"] Oct 09 10:12:52 crc kubenswrapper[4923]: I1009 10:12:52.610966 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02a21759-421a-429b-ac1b-c43cd7c236f0" path="/var/lib/kubelet/pods/02a21759-421a-429b-ac1b-c43cd7c236f0/volumes" Oct 09 10:14:24 crc kubenswrapper[4923]: I1009 10:14:24.599499 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:14:24 crc kubenswrapper[4923]: I1009 10:14:24.600149 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:14:54 crc kubenswrapper[4923]: I1009 10:14:54.600461 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:14:54 crc kubenswrapper[4923]: I1009 10:14:54.601498 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.146863 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9"] Oct 09 10:15:00 crc kubenswrapper[4923]: E1009 10:15:00.147569 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02a21759-421a-429b-ac1b-c43cd7c236f0" containerName="registry" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.147586 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="02a21759-421a-429b-ac1b-c43cd7c236f0" containerName="registry" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.147695 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="02a21759-421a-429b-ac1b-c43cd7c236f0" containerName="registry" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.148163 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.152048 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.152691 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.164115 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9"] Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.262551 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42r44\" (UniqueName: \"kubernetes.io/projected/09667cd2-7829-459b-a8bb-0cd255c221ab-kube-api-access-42r44\") pod \"collect-profiles-29333415-h4fp9\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.262643 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09667cd2-7829-459b-a8bb-0cd255c221ab-config-volume\") pod \"collect-profiles-29333415-h4fp9\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.262667 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09667cd2-7829-459b-a8bb-0cd255c221ab-secret-volume\") pod \"collect-profiles-29333415-h4fp9\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.363960 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09667cd2-7829-459b-a8bb-0cd255c221ab-config-volume\") pod \"collect-profiles-29333415-h4fp9\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.364385 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09667cd2-7829-459b-a8bb-0cd255c221ab-secret-volume\") pod \"collect-profiles-29333415-h4fp9\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.364609 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42r44\" (UniqueName: \"kubernetes.io/projected/09667cd2-7829-459b-a8bb-0cd255c221ab-kube-api-access-42r44\") pod \"collect-profiles-29333415-h4fp9\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.365140 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09667cd2-7829-459b-a8bb-0cd255c221ab-config-volume\") pod \"collect-profiles-29333415-h4fp9\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.373846 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09667cd2-7829-459b-a8bb-0cd255c221ab-secret-volume\") pod \"collect-profiles-29333415-h4fp9\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.393038 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42r44\" (UniqueName: \"kubernetes.io/projected/09667cd2-7829-459b-a8bb-0cd255c221ab-kube-api-access-42r44\") pod \"collect-profiles-29333415-h4fp9\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.482852 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.690825 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9"] Oct 09 10:15:00 crc kubenswrapper[4923]: I1009 10:15:00.707581 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" event={"ID":"09667cd2-7829-459b-a8bb-0cd255c221ab","Type":"ContainerStarted","Data":"d4b1540c4170b1f6db7d4caeb5ae19540fa3b64463faf119603dc80cc63f42c3"} Oct 09 10:15:01 crc kubenswrapper[4923]: I1009 10:15:01.715404 4923 generic.go:334] "Generic (PLEG): container finished" podID="09667cd2-7829-459b-a8bb-0cd255c221ab" containerID="093c9f9c8a360e261fdfe3790d91daac15d0d48833a1881979182167619b3917" exitCode=0 Oct 09 10:15:01 crc kubenswrapper[4923]: I1009 10:15:01.715466 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" event={"ID":"09667cd2-7829-459b-a8bb-0cd255c221ab","Type":"ContainerDied","Data":"093c9f9c8a360e261fdfe3790d91daac15d0d48833a1881979182167619b3917"} Oct 09 10:15:02 crc kubenswrapper[4923]: I1009 10:15:02.923919 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.097968 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42r44\" (UniqueName: \"kubernetes.io/projected/09667cd2-7829-459b-a8bb-0cd255c221ab-kube-api-access-42r44\") pod \"09667cd2-7829-459b-a8bb-0cd255c221ab\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.098070 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09667cd2-7829-459b-a8bb-0cd255c221ab-secret-volume\") pod \"09667cd2-7829-459b-a8bb-0cd255c221ab\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.098130 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09667cd2-7829-459b-a8bb-0cd255c221ab-config-volume\") pod \"09667cd2-7829-459b-a8bb-0cd255c221ab\" (UID: \"09667cd2-7829-459b-a8bb-0cd255c221ab\") " Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.099215 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09667cd2-7829-459b-a8bb-0cd255c221ab-config-volume" (OuterVolumeSpecName: "config-volume") pod "09667cd2-7829-459b-a8bb-0cd255c221ab" (UID: "09667cd2-7829-459b-a8bb-0cd255c221ab"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.104051 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09667cd2-7829-459b-a8bb-0cd255c221ab-kube-api-access-42r44" (OuterVolumeSpecName: "kube-api-access-42r44") pod "09667cd2-7829-459b-a8bb-0cd255c221ab" (UID: "09667cd2-7829-459b-a8bb-0cd255c221ab"). InnerVolumeSpecName "kube-api-access-42r44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.104198 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09667cd2-7829-459b-a8bb-0cd255c221ab-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "09667cd2-7829-459b-a8bb-0cd255c221ab" (UID: "09667cd2-7829-459b-a8bb-0cd255c221ab"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.199832 4923 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/09667cd2-7829-459b-a8bb-0cd255c221ab-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.199864 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/09667cd2-7829-459b-a8bb-0cd255c221ab-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.199874 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42r44\" (UniqueName: \"kubernetes.io/projected/09667cd2-7829-459b-a8bb-0cd255c221ab-kube-api-access-42r44\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.728663 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" event={"ID":"09667cd2-7829-459b-a8bb-0cd255c221ab","Type":"ContainerDied","Data":"d4b1540c4170b1f6db7d4caeb5ae19540fa3b64463faf119603dc80cc63f42c3"} Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.728723 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4b1540c4170b1f6db7d4caeb5ae19540fa3b64463faf119603dc80cc63f42c3" Oct 09 10:15:03 crc kubenswrapper[4923]: I1009 10:15:03.728817 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9" Oct 09 10:15:24 crc kubenswrapper[4923]: I1009 10:15:24.599691 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:15:24 crc kubenswrapper[4923]: I1009 10:15:24.600712 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:15:24 crc kubenswrapper[4923]: I1009 10:15:24.611079 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:15:24 crc kubenswrapper[4923]: I1009 10:15:24.611984 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"15e3211b265dc1c111dcb9faf998805ee133fefab85afa177cdecf472b4fa20c"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:15:24 crc kubenswrapper[4923]: I1009 10:15:24.612102 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://15e3211b265dc1c111dcb9faf998805ee133fefab85afa177cdecf472b4fa20c" gracePeriod=600 Oct 09 10:15:24 crc kubenswrapper[4923]: I1009 10:15:24.874109 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="15e3211b265dc1c111dcb9faf998805ee133fefab85afa177cdecf472b4fa20c" exitCode=0 Oct 09 10:15:24 crc kubenswrapper[4923]: I1009 10:15:24.874222 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"15e3211b265dc1c111dcb9faf998805ee133fefab85afa177cdecf472b4fa20c"} Oct 09 10:15:24 crc kubenswrapper[4923]: I1009 10:15:24.874534 4923 scope.go:117] "RemoveContainer" containerID="fdc7d576a4e21a1ee542a4eba76ea241d8e56b0e66e0f288027bd3d2de345b25" Oct 09 10:15:25 crc kubenswrapper[4923]: I1009 10:15:25.886108 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"af930c04f923630accaccaa4b93be2a9110fb2ade8cd00a8a79620f54fdb4120"} Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.586945 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-kclss"] Oct 09 10:15:27 crc kubenswrapper[4923]: E1009 10:15:27.588040 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09667cd2-7829-459b-a8bb-0cd255c221ab" containerName="collect-profiles" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.588079 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="09667cd2-7829-459b-a8bb-0cd255c221ab" containerName="collect-profiles" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.588282 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="09667cd2-7829-459b-a8bb-0cd255c221ab" containerName="collect-profiles" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.589167 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-kclss" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.591713 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.591894 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-wsrcj" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.591961 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.598422 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-tfrn9"] Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.599548 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-tfrn9" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.602017 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-s22m7" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.605677 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xhmhr"] Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.612829 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-kclss"] Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.613097 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.618155 4923 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-kvdpf" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.618874 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xhmhr"] Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.622581 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-tfrn9"] Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.650317 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj9t9\" (UniqueName: \"kubernetes.io/projected/46e74a19-0dfc-4cac-a8ba-98062a9ca3ed-kube-api-access-xj9t9\") pod \"cert-manager-5b446d88c5-kclss\" (UID: \"46e74a19-0dfc-4cac-a8ba-98062a9ca3ed\") " pod="cert-manager/cert-manager-5b446d88c5-kclss" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.752896 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zdj8\" (UniqueName: \"kubernetes.io/projected/5155ab6c-6e6c-4937-85cd-55fc49b7ce1a-kube-api-access-6zdj8\") pod \"cert-manager-webhook-5655c58dd6-xhmhr\" (UID: \"5155ab6c-6e6c-4937-85cd-55fc49b7ce1a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.753459 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjpc2\" (UniqueName: \"kubernetes.io/projected/9aac563b-4a10-4a9e-a5c0-b3dfb5e4476d-kube-api-access-pjpc2\") pod \"cert-manager-cainjector-7f985d654d-tfrn9\" (UID: \"9aac563b-4a10-4a9e-a5c0-b3dfb5e4476d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-tfrn9" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.753501 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj9t9\" (UniqueName: \"kubernetes.io/projected/46e74a19-0dfc-4cac-a8ba-98062a9ca3ed-kube-api-access-xj9t9\") pod \"cert-manager-5b446d88c5-kclss\" (UID: \"46e74a19-0dfc-4cac-a8ba-98062a9ca3ed\") " pod="cert-manager/cert-manager-5b446d88c5-kclss" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.787955 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj9t9\" (UniqueName: \"kubernetes.io/projected/46e74a19-0dfc-4cac-a8ba-98062a9ca3ed-kube-api-access-xj9t9\") pod \"cert-manager-5b446d88c5-kclss\" (UID: \"46e74a19-0dfc-4cac-a8ba-98062a9ca3ed\") " pod="cert-manager/cert-manager-5b446d88c5-kclss" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.854047 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjpc2\" (UniqueName: \"kubernetes.io/projected/9aac563b-4a10-4a9e-a5c0-b3dfb5e4476d-kube-api-access-pjpc2\") pod \"cert-manager-cainjector-7f985d654d-tfrn9\" (UID: \"9aac563b-4a10-4a9e-a5c0-b3dfb5e4476d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-tfrn9" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.854157 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zdj8\" (UniqueName: \"kubernetes.io/projected/5155ab6c-6e6c-4937-85cd-55fc49b7ce1a-kube-api-access-6zdj8\") pod \"cert-manager-webhook-5655c58dd6-xhmhr\" (UID: \"5155ab6c-6e6c-4937-85cd-55fc49b7ce1a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.872902 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zdj8\" (UniqueName: \"kubernetes.io/projected/5155ab6c-6e6c-4937-85cd-55fc49b7ce1a-kube-api-access-6zdj8\") pod \"cert-manager-webhook-5655c58dd6-xhmhr\" (UID: \"5155ab6c-6e6c-4937-85cd-55fc49b7ce1a\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.876859 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjpc2\" (UniqueName: \"kubernetes.io/projected/9aac563b-4a10-4a9e-a5c0-b3dfb5e4476d-kube-api-access-pjpc2\") pod \"cert-manager-cainjector-7f985d654d-tfrn9\" (UID: \"9aac563b-4a10-4a9e-a5c0-b3dfb5e4476d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-tfrn9" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.918060 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-kclss" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.927740 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-tfrn9" Oct 09 10:15:27 crc kubenswrapper[4923]: I1009 10:15:27.952221 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" Oct 09 10:15:28 crc kubenswrapper[4923]: I1009 10:15:28.199274 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-kclss"] Oct 09 10:15:28 crc kubenswrapper[4923]: W1009 10:15:28.211925 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46e74a19_0dfc_4cac_a8ba_98062a9ca3ed.slice/crio-91a164cb5fc704a66893917d3c4d4f85bf3906fbc109b9b94577018bba4eba2b WatchSource:0}: Error finding container 91a164cb5fc704a66893917d3c4d4f85bf3906fbc109b9b94577018bba4eba2b: Status 404 returned error can't find the container with id 91a164cb5fc704a66893917d3c4d4f85bf3906fbc109b9b94577018bba4eba2b Oct 09 10:15:28 crc kubenswrapper[4923]: I1009 10:15:28.220890 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 10:15:28 crc kubenswrapper[4923]: I1009 10:15:28.251589 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-tfrn9"] Oct 09 10:15:28 crc kubenswrapper[4923]: W1009 10:15:28.255854 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9aac563b_4a10_4a9e_a5c0_b3dfb5e4476d.slice/crio-be195d2f9dfef33c967cd8b83849407400c02d19130423c83c59ae6bd5234faa WatchSource:0}: Error finding container be195d2f9dfef33c967cd8b83849407400c02d19130423c83c59ae6bd5234faa: Status 404 returned error can't find the container with id be195d2f9dfef33c967cd8b83849407400c02d19130423c83c59ae6bd5234faa Oct 09 10:15:28 crc kubenswrapper[4923]: I1009 10:15:28.274633 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xhmhr"] Oct 09 10:15:28 crc kubenswrapper[4923]: W1009 10:15:28.279089 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5155ab6c_6e6c_4937_85cd_55fc49b7ce1a.slice/crio-b4bb2ed0c8cb029c75b1e171e38c473b17439916cf00cd8890138121564bf5e3 WatchSource:0}: Error finding container b4bb2ed0c8cb029c75b1e171e38c473b17439916cf00cd8890138121564bf5e3: Status 404 returned error can't find the container with id b4bb2ed0c8cb029c75b1e171e38c473b17439916cf00cd8890138121564bf5e3 Oct 09 10:15:28 crc kubenswrapper[4923]: I1009 10:15:28.906596 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-tfrn9" event={"ID":"9aac563b-4a10-4a9e-a5c0-b3dfb5e4476d","Type":"ContainerStarted","Data":"be195d2f9dfef33c967cd8b83849407400c02d19130423c83c59ae6bd5234faa"} Oct 09 10:15:28 crc kubenswrapper[4923]: I1009 10:15:28.909183 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" event={"ID":"5155ab6c-6e6c-4937-85cd-55fc49b7ce1a","Type":"ContainerStarted","Data":"b4bb2ed0c8cb029c75b1e171e38c473b17439916cf00cd8890138121564bf5e3"} Oct 09 10:15:28 crc kubenswrapper[4923]: I1009 10:15:28.910958 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-kclss" event={"ID":"46e74a19-0dfc-4cac-a8ba-98062a9ca3ed","Type":"ContainerStarted","Data":"91a164cb5fc704a66893917d3c4d4f85bf3906fbc109b9b94577018bba4eba2b"} Oct 09 10:15:31 crc kubenswrapper[4923]: I1009 10:15:31.942336 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-tfrn9" event={"ID":"9aac563b-4a10-4a9e-a5c0-b3dfb5e4476d","Type":"ContainerStarted","Data":"1adf0f2f278cef251d235f8c8f2efe9cf5469110da3bfa6d8293102c036f702c"} Oct 09 10:15:31 crc kubenswrapper[4923]: I1009 10:15:31.943870 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" event={"ID":"5155ab6c-6e6c-4937-85cd-55fc49b7ce1a","Type":"ContainerStarted","Data":"ce6c8b64bae26a5b8195ad602e959a2a000cb82f4a216c503af452c5ca371aa5"} Oct 09 10:15:31 crc kubenswrapper[4923]: I1009 10:15:31.944123 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" Oct 09 10:15:31 crc kubenswrapper[4923]: I1009 10:15:31.946649 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-kclss" event={"ID":"46e74a19-0dfc-4cac-a8ba-98062a9ca3ed","Type":"ContainerStarted","Data":"49c25fbaa30766becb95ef5622a614802935f31c8e77e8845119d3b287ca7fac"} Oct 09 10:15:31 crc kubenswrapper[4923]: I1009 10:15:31.988878 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-tfrn9" podStartSLOduration=2.353808345 podStartE2EDuration="4.988853776s" podCreationTimestamp="2025-10-09 10:15:27 +0000 UTC" firstStartedPulling="2025-10-09 10:15:28.25948283 +0000 UTC m=+614.327664586" lastFinishedPulling="2025-10-09 10:15:30.894528261 +0000 UTC m=+616.962710017" observedRunningTime="2025-10-09 10:15:31.966444928 +0000 UTC m=+618.034626684" watchObservedRunningTime="2025-10-09 10:15:31.988853776 +0000 UTC m=+618.057035532" Oct 09 10:15:32 crc kubenswrapper[4923]: I1009 10:15:32.108968 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-kclss" podStartSLOduration=1.70341553 podStartE2EDuration="5.108938121s" podCreationTimestamp="2025-10-09 10:15:27 +0000 UTC" firstStartedPulling="2025-10-09 10:15:28.22057704 +0000 UTC m=+614.288758796" lastFinishedPulling="2025-10-09 10:15:31.626099631 +0000 UTC m=+617.694281387" observedRunningTime="2025-10-09 10:15:31.989607838 +0000 UTC m=+618.057789594" watchObservedRunningTime="2025-10-09 10:15:32.108938121 +0000 UTC m=+618.177119877" Oct 09 10:15:37 crc kubenswrapper[4923]: I1009 10:15:37.957323 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" Oct 09 10:15:37 crc kubenswrapper[4923]: I1009 10:15:37.989088 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-xhmhr" podStartSLOduration=7.713348374 podStartE2EDuration="10.989052606s" podCreationTimestamp="2025-10-09 10:15:27 +0000 UTC" firstStartedPulling="2025-10-09 10:15:28.281864218 +0000 UTC m=+614.350045984" lastFinishedPulling="2025-10-09 10:15:31.55756846 +0000 UTC m=+617.625750216" observedRunningTime="2025-10-09 10:15:32.113969922 +0000 UTC m=+618.182151688" watchObservedRunningTime="2025-10-09 10:15:37.989052606 +0000 UTC m=+624.057234402" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.065300 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-75qcv"] Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.065672 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovn-controller" containerID="cri-o://dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f" gracePeriod=30 Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.066041 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b" gracePeriod=30 Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.066086 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="northd" containerID="cri-o://f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034" gracePeriod=30 Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.066156 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kube-rbac-proxy-node" containerID="cri-o://f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4" gracePeriod=30 Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.066256 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="sbdb" containerID="cri-o://5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f" gracePeriod=30 Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.066348 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="nbdb" containerID="cri-o://e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9" gracePeriod=30 Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.066105 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovn-acl-logging" containerID="cri-o://0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5" gracePeriod=30 Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.134568 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" containerID="cri-o://f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" gracePeriod=30 Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.270044 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 is running failed: container process not found" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.270621 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 is running failed: container process not found" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.271104 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 is running failed: container process not found" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" cmd=["/bin/bash","-c","#!/bin/bash\ntest -f /etc/cni/net.d/10-ovn-kubernetes.conf\n"] Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.271144 4923 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.429587 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/3.log" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.433021 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovn-acl-logging/0.log" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.433628 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovn-controller/0.log" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.434196 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.521632 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-tp98r"] Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522068 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovn-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522098 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovn-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522118 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kube-rbac-proxy-ovn-metrics" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522134 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kube-rbac-proxy-ovn-metrics" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522162 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovn-acl-logging" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522177 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovn-acl-logging" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522193 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kubecfg-setup" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522210 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kubecfg-setup" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522232 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522245 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522263 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="northd" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522277 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="northd" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522297 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522310 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522340 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522354 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522371 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522384 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522401 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="sbdb" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522415 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="sbdb" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522440 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="nbdb" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522455 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="nbdb" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522471 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kube-rbac-proxy-node" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522485 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kube-rbac-proxy-node" Oct 09 10:15:38 crc kubenswrapper[4923]: E1009 10:15:38.522507 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522520 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522697 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522719 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="northd" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522734 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovn-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522788 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovn-acl-logging" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522816 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kube-rbac-proxy-node" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522832 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="nbdb" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522855 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="sbdb" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522870 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522884 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522902 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.522921 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="kube-rbac-proxy-ovn-metrics" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.523257 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc2a175c-508c-4f50-8205-013a9482c989" containerName="ovnkube-controller" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.527332 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615106 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-systemd-units\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615205 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-ovn-kubernetes\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615251 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-log-socket\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615299 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-etc-openvswitch\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615351 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-systemd\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615295 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615333 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-log-socket" (OuterVolumeSpecName: "log-socket") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615391 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615390 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615434 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-var-lib-cni-networks-ovn-kubernetes\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615495 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615852 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-netd\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.615915 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.616185 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-bin\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.616227 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.616495 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v9j6\" (UniqueName: \"kubernetes.io/projected/bc2a175c-508c-4f50-8205-013a9482c989-kube-api-access-8v9j6\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.618614 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-ovn\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.618697 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.618920 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-config\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.619359 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc2a175c-508c-4f50-8205-013a9482c989-ovn-node-metrics-cert\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.619794 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-kubelet\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.619857 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-var-lib-openvswitch\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.619942 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.619951 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.619998 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.620034 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-script-lib\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.620076 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-node-log\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.620122 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-env-overrides\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.620156 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-netns\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.620185 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-openvswitch\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.620257 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-slash" (OuterVolumeSpecName: "host-slash") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.620214 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-slash\") pod \"bc2a175c-508c-4f50-8205-013a9482c989\" (UID: \"bc2a175c-508c-4f50-8205-013a9482c989\") " Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.621191 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-node-log" (OuterVolumeSpecName: "node-log") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.621310 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.621449 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.622231 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-run-ovn\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.622291 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f0d84dc9-d157-41c4-a85e-1daba23b31ce-ovnkube-script-lib\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.622313 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-log-socket\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.622447 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.622584 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6kqs\" (UniqueName: \"kubernetes.io/projected/f0d84dc9-d157-41c4-a85e-1daba23b31ce-kube-api-access-w6kqs\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.622901 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-var-lib-openvswitch\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.622912 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.623010 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-run-systemd\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.623374 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-run-openvswitch\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.623483 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-kubelet\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.623604 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f0d84dc9-d157-41c4-a85e-1daba23b31ce-ovnkube-config\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.623670 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.623863 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-cni-bin\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.623981 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-run-ovn-kubernetes\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.624066 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f0d84dc9-d157-41c4-a85e-1daba23b31ce-env-overrides\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.624118 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-node-log\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.624193 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f0d84dc9-d157-41c4-a85e-1daba23b31ce-ovn-node-metrics-cert\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.624243 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-slash\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.624410 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-run-netns\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.624532 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-etc-openvswitch\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.624597 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-systemd-units\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.624644 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-cni-netd\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625115 4923 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625162 4923 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625191 4923 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625216 4923 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625238 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625264 4923 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625292 4923 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625318 4923 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625341 4923 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-node-log\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625367 4923 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bc2a175c-508c-4f50-8205-013a9482c989-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625393 4923 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625416 4923 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625441 4923 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-slash\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625466 4923 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625495 4923 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625525 4923 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-log-socket\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.625552 4923 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.626537 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc2a175c-508c-4f50-8205-013a9482c989-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.626645 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc2a175c-508c-4f50-8205-013a9482c989-kube-api-access-8v9j6" (OuterVolumeSpecName: "kube-api-access-8v9j6") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "kube-api-access-8v9j6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.647678 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "bc2a175c-508c-4f50-8205-013a9482c989" (UID: "bc2a175c-508c-4f50-8205-013a9482c989"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727201 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-run-netns\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727293 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-etc-openvswitch\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727333 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-systemd-units\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727372 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-cni-netd\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727424 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-run-ovn\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727436 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-etc-openvswitch\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727540 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-systemd-units\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727438 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-run-netns\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727474 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f0d84dc9-d157-41c4-a85e-1daba23b31ce-ovnkube-script-lib\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727674 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-log-socket\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727730 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6kqs\" (UniqueName: \"kubernetes.io/projected/f0d84dc9-d157-41c4-a85e-1daba23b31ce-kube-api-access-w6kqs\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727744 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-log-socket\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727880 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-var-lib-openvswitch\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727934 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-run-systemd\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727946 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-var-lib-openvswitch\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727519 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-cni-netd\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728021 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-run-systemd\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728040 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-run-openvswitch\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727989 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-run-openvswitch\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728166 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-kubelet\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728254 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f0d84dc9-d157-41c4-a85e-1daba23b31ce-ovnkube-config\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728269 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-kubelet\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728320 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728372 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728425 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-cni-bin\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.727519 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-run-ovn\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728500 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-run-ovn-kubernetes\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728553 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-cni-bin\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728573 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f0d84dc9-d157-41c4-a85e-1daba23b31ce-env-overrides\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728591 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-run-ovn-kubernetes\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728629 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-node-log\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728710 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f0d84dc9-d157-41c4-a85e-1daba23b31ce-ovn-node-metrics-cert\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728808 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-slash\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728838 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-node-log\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728881 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f0d84dc9-d157-41c4-a85e-1daba23b31ce-host-slash\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728930 4923 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/bc2a175c-508c-4f50-8205-013a9482c989-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.728975 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v9j6\" (UniqueName: \"kubernetes.io/projected/bc2a175c-508c-4f50-8205-013a9482c989-kube-api-access-8v9j6\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.729008 4923 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bc2a175c-508c-4f50-8205-013a9482c989-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.729863 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f0d84dc9-d157-41c4-a85e-1daba23b31ce-env-overrides\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.729884 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f0d84dc9-d157-41c4-a85e-1daba23b31ce-ovnkube-config\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.730061 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f0d84dc9-d157-41c4-a85e-1daba23b31ce-ovnkube-script-lib\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.733863 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f0d84dc9-d157-41c4-a85e-1daba23b31ce-ovn-node-metrics-cert\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.752386 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6kqs\" (UniqueName: \"kubernetes.io/projected/f0d84dc9-d157-41c4-a85e-1daba23b31ce-kube-api-access-w6kqs\") pod \"ovnkube-node-tp98r\" (UID: \"f0d84dc9-d157-41c4-a85e-1daba23b31ce\") " pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:38 crc kubenswrapper[4923]: I1009 10:15:38.847642 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.007324 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerStarted","Data":"99ee3bf24192e857a13339135a54915c3bbdb5dc8f6237474c906ff46366369d"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.011516 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovnkube-controller/3.log" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.014916 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovn-acl-logging/0.log" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.015734 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-75qcv_bc2a175c-508c-4f50-8205-013a9482c989/ovn-controller/0.log" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016290 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" exitCode=0 Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016404 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f" exitCode=0 Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016355 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016510 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016528 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016548 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016573 4923 scope.go:117] "RemoveContainer" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016485 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9" exitCode=0 Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016652 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034" exitCode=0 Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016685 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b" exitCode=0 Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016701 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4" exitCode=0 Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016718 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016736 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016778 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016803 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016820 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016828 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016836 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016854 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016863 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016875 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016884 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016893 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016904 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016720 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5" exitCode=143 Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016946 4923 generic.go:334] "Generic (PLEG): container finished" podID="bc2a175c-508c-4f50-8205-013a9482c989" containerID="dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f" exitCode=143 Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.016920 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017013 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017024 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017031 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017037 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017044 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017050 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017057 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017064 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017070 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017081 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017095 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017103 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017109 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017115 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017121 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017127 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017133 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017139 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017147 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017154 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017164 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-75qcv" event={"ID":"bc2a175c-508c-4f50-8205-013a9482c989","Type":"ContainerDied","Data":"6e6a1e686208e721282523d9ce50dd44ac2b857ab0c9e47061a12222cf97e75f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017180 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017190 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017200 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017207 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017215 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017224 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017232 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017240 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017248 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.017255 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.021136 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/2.log" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.022437 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/1.log" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.022582 4923 generic.go:334] "Generic (PLEG): container finished" podID="d5c5f455-4e94-4f9a-b6fa-2535685c6354" containerID="8aa38bf09b127046fbe3053995133ff9c81ee52e4a4cd7e17444e49b1e6ce556" exitCode=2 Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.022692 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4cn4f" event={"ID":"d5c5f455-4e94-4f9a-b6fa-2535685c6354","Type":"ContainerDied","Data":"8aa38bf09b127046fbe3053995133ff9c81ee52e4a4cd7e17444e49b1e6ce556"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.022786 4923 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a"} Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.023538 4923 scope.go:117] "RemoveContainer" containerID="8aa38bf09b127046fbe3053995133ff9c81ee52e4a4cd7e17444e49b1e6ce556" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.023909 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4cn4f_openshift-multus(d5c5f455-4e94-4f9a-b6fa-2535685c6354)\"" pod="openshift-multus/multus-4cn4f" podUID="d5c5f455-4e94-4f9a-b6fa-2535685c6354" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.079147 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.106829 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-75qcv"] Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.107962 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-75qcv"] Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.110637 4923 scope.go:117] "RemoveContainer" containerID="5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.131541 4923 scope.go:117] "RemoveContainer" containerID="e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.147802 4923 scope.go:117] "RemoveContainer" containerID="f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.162630 4923 scope.go:117] "RemoveContainer" containerID="99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.177131 4923 scope.go:117] "RemoveContainer" containerID="f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.192234 4923 scope.go:117] "RemoveContainer" containerID="0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.251435 4923 scope.go:117] "RemoveContainer" containerID="dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.266639 4923 scope.go:117] "RemoveContainer" containerID="657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.282470 4923 scope.go:117] "RemoveContainer" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.283023 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": container with ID starting with f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 not found: ID does not exist" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.283073 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} err="failed to get container status \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": rpc error: code = NotFound desc = could not find container \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": container with ID starting with f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.283109 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.283455 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\": container with ID starting with 013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd not found: ID does not exist" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.283491 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd"} err="failed to get container status \"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\": rpc error: code = NotFound desc = could not find container \"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\": container with ID starting with 013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.283508 4923 scope.go:117] "RemoveContainer" containerID="5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.283876 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\": container with ID starting with 5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f not found: ID does not exist" containerID="5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.283919 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} err="failed to get container status \"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\": rpc error: code = NotFound desc = could not find container \"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\": container with ID starting with 5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.283941 4923 scope.go:117] "RemoveContainer" containerID="e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.284291 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\": container with ID starting with e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9 not found: ID does not exist" containerID="e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.284325 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} err="failed to get container status \"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\": rpc error: code = NotFound desc = could not find container \"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\": container with ID starting with e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.284344 4923 scope.go:117] "RemoveContainer" containerID="f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.284956 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\": container with ID starting with f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034 not found: ID does not exist" containerID="f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.285025 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} err="failed to get container status \"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\": rpc error: code = NotFound desc = could not find container \"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\": container with ID starting with f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.285082 4923 scope.go:117] "RemoveContainer" containerID="99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.285481 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\": container with ID starting with 99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b not found: ID does not exist" containerID="99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.285505 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} err="failed to get container status \"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\": rpc error: code = NotFound desc = could not find container \"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\": container with ID starting with 99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.285524 4923 scope.go:117] "RemoveContainer" containerID="f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.285878 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\": container with ID starting with f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4 not found: ID does not exist" containerID="f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.285913 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} err="failed to get container status \"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\": rpc error: code = NotFound desc = could not find container \"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\": container with ID starting with f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.285935 4923 scope.go:117] "RemoveContainer" containerID="0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.286266 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\": container with ID starting with 0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5 not found: ID does not exist" containerID="0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.286294 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} err="failed to get container status \"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\": rpc error: code = NotFound desc = could not find container \"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\": container with ID starting with 0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.286310 4923 scope.go:117] "RemoveContainer" containerID="dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.286649 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\": container with ID starting with dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f not found: ID does not exist" containerID="dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.286676 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} err="failed to get container status \"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\": rpc error: code = NotFound desc = could not find container \"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\": container with ID starting with dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.286690 4923 scope.go:117] "RemoveContainer" containerID="657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b" Oct 09 10:15:39 crc kubenswrapper[4923]: E1009 10:15:39.287101 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\": container with ID starting with 657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b not found: ID does not exist" containerID="657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.287143 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b"} err="failed to get container status \"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\": rpc error: code = NotFound desc = could not find container \"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\": container with ID starting with 657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.287173 4923 scope.go:117] "RemoveContainer" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.287538 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} err="failed to get container status \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": rpc error: code = NotFound desc = could not find container \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": container with ID starting with f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.287566 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.287920 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd"} err="failed to get container status \"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\": rpc error: code = NotFound desc = could not find container \"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\": container with ID starting with 013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.287956 4923 scope.go:117] "RemoveContainer" containerID="5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.288386 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} err="failed to get container status \"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\": rpc error: code = NotFound desc = could not find container \"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\": container with ID starting with 5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.288416 4923 scope.go:117] "RemoveContainer" containerID="e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.288829 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} err="failed to get container status \"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\": rpc error: code = NotFound desc = could not find container \"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\": container with ID starting with e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.288859 4923 scope.go:117] "RemoveContainer" containerID="f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.289231 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} err="failed to get container status \"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\": rpc error: code = NotFound desc = could not find container \"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\": container with ID starting with f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.289269 4923 scope.go:117] "RemoveContainer" containerID="99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.289523 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} err="failed to get container status \"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\": rpc error: code = NotFound desc = could not find container \"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\": container with ID starting with 99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.289563 4923 scope.go:117] "RemoveContainer" containerID="f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.289976 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} err="failed to get container status \"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\": rpc error: code = NotFound desc = could not find container \"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\": container with ID starting with f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.290008 4923 scope.go:117] "RemoveContainer" containerID="0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.290294 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} err="failed to get container status \"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\": rpc error: code = NotFound desc = could not find container \"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\": container with ID starting with 0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.290335 4923 scope.go:117] "RemoveContainer" containerID="dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.290867 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} err="failed to get container status \"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\": rpc error: code = NotFound desc = could not find container \"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\": container with ID starting with dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.290896 4923 scope.go:117] "RemoveContainer" containerID="657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.291294 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b"} err="failed to get container status \"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\": rpc error: code = NotFound desc = could not find container \"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\": container with ID starting with 657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.291331 4923 scope.go:117] "RemoveContainer" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.291615 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} err="failed to get container status \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": rpc error: code = NotFound desc = could not find container \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": container with ID starting with f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.291642 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.292096 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd"} err="failed to get container status \"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\": rpc error: code = NotFound desc = could not find container \"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\": container with ID starting with 013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.292160 4923 scope.go:117] "RemoveContainer" containerID="5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.292469 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} err="failed to get container status \"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\": rpc error: code = NotFound desc = could not find container \"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\": container with ID starting with 5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.292501 4923 scope.go:117] "RemoveContainer" containerID="e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.292819 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} err="failed to get container status \"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\": rpc error: code = NotFound desc = could not find container \"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\": container with ID starting with e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.292858 4923 scope.go:117] "RemoveContainer" containerID="f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.293156 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} err="failed to get container status \"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\": rpc error: code = NotFound desc = could not find container \"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\": container with ID starting with f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.293193 4923 scope.go:117] "RemoveContainer" containerID="99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.293428 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} err="failed to get container status \"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\": rpc error: code = NotFound desc = could not find container \"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\": container with ID starting with 99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.293455 4923 scope.go:117] "RemoveContainer" containerID="f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.293697 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} err="failed to get container status \"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\": rpc error: code = NotFound desc = could not find container \"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\": container with ID starting with f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.293719 4923 scope.go:117] "RemoveContainer" containerID="0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.294060 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} err="failed to get container status \"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\": rpc error: code = NotFound desc = could not find container \"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\": container with ID starting with 0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.294086 4923 scope.go:117] "RemoveContainer" containerID="dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.294353 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} err="failed to get container status \"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\": rpc error: code = NotFound desc = could not find container \"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\": container with ID starting with dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.294385 4923 scope.go:117] "RemoveContainer" containerID="657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.294676 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b"} err="failed to get container status \"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\": rpc error: code = NotFound desc = could not find container \"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\": container with ID starting with 657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.294705 4923 scope.go:117] "RemoveContainer" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.295019 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} err="failed to get container status \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": rpc error: code = NotFound desc = could not find container \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": container with ID starting with f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.295058 4923 scope.go:117] "RemoveContainer" containerID="013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.295476 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd"} err="failed to get container status \"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\": rpc error: code = NotFound desc = could not find container \"013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd\": container with ID starting with 013de9dd0b4d81c94e326b7009756324341866e21c7fb6643b35122a69a4cadd not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.295502 4923 scope.go:117] "RemoveContainer" containerID="5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.295740 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f"} err="failed to get container status \"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\": rpc error: code = NotFound desc = could not find container \"5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f\": container with ID starting with 5ae9e5527416d461e78ccafddc3d77d5f4c41ddd2ff8855750348872968b5d5f not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.295773 4923 scope.go:117] "RemoveContainer" containerID="e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.295990 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9"} err="failed to get container status \"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\": rpc error: code = NotFound desc = could not find container \"e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9\": container with ID starting with e6d2ac65352e5e74395c31c8f8133f21ba3696625fdc68ce39b0ec10ef8312a9 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.296012 4923 scope.go:117] "RemoveContainer" containerID="f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.296354 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034"} err="failed to get container status \"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\": rpc error: code = NotFound desc = could not find container \"f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034\": container with ID starting with f2ba3ad75240b8d12442239ab6a992996c51c9cca7e092a82ec1fc4b9ea33034 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.296375 4923 scope.go:117] "RemoveContainer" containerID="99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.296584 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b"} err="failed to get container status \"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\": rpc error: code = NotFound desc = could not find container \"99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b\": container with ID starting with 99dcbb4b15492cdc8f2b3dc98ce89661561473a2815e89c5fadebd600af7417b not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.296614 4923 scope.go:117] "RemoveContainer" containerID="f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.296971 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4"} err="failed to get container status \"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\": rpc error: code = NotFound desc = could not find container \"f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4\": container with ID starting with f9561eccf3101fe22f32df19986d8dd1fdb8b71b4668e13a0f51c34faad06ac4 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.296993 4923 scope.go:117] "RemoveContainer" containerID="0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.297180 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5"} err="failed to get container status \"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\": rpc error: code = NotFound desc = could not find container \"0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5\": container with ID starting with 0dd1d168fed5a84ead0cecf0666e073b280d5edeaa8c9ce232560aa7053c42f5 not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.297205 4923 scope.go:117] "RemoveContainer" containerID="dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.297413 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f"} err="failed to get container status \"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\": rpc error: code = NotFound desc = could not find container \"dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f\": container with ID starting with dc974d9059df523043f45d78a7526008262b25f929e186cc1db539f611a0299f not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.297438 4923 scope.go:117] "RemoveContainer" containerID="657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.297687 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b"} err="failed to get container status \"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\": rpc error: code = NotFound desc = could not find container \"657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b\": container with ID starting with 657ce7a7ad9e92ce474b8c63d9d24dd432503d78b8cffe349bbbfc4cf1eb5a2b not found: ID does not exist" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.297707 4923 scope.go:117] "RemoveContainer" containerID="f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0" Oct 09 10:15:39 crc kubenswrapper[4923]: I1009 10:15:39.297952 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0"} err="failed to get container status \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": rpc error: code = NotFound desc = could not find container \"f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0\": container with ID starting with f999592a568f70ed2ff8039cd1f6806e64763f9cef521d8f143e26c1a0c099f0 not found: ID does not exist" Oct 09 10:15:40 crc kubenswrapper[4923]: I1009 10:15:40.034589 4923 generic.go:334] "Generic (PLEG): container finished" podID="f0d84dc9-d157-41c4-a85e-1daba23b31ce" containerID="f3a9b2d96d82ceb2c2fe4fb2e5501d7ff72f92dfb9b6fde6c0e46413c13f08ed" exitCode=0 Oct 09 10:15:40 crc kubenswrapper[4923]: I1009 10:15:40.034706 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerDied","Data":"f3a9b2d96d82ceb2c2fe4fb2e5501d7ff72f92dfb9b6fde6c0e46413c13f08ed"} Oct 09 10:15:40 crc kubenswrapper[4923]: I1009 10:15:40.610930 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc2a175c-508c-4f50-8205-013a9482c989" path="/var/lib/kubelet/pods/bc2a175c-508c-4f50-8205-013a9482c989/volumes" Oct 09 10:15:41 crc kubenswrapper[4923]: I1009 10:15:41.045997 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerStarted","Data":"5c5627a54149806ded5457bbc4034a2776a279c50b590e175b29e8da7739fb4b"} Oct 09 10:15:41 crc kubenswrapper[4923]: I1009 10:15:41.046048 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerStarted","Data":"5f751aaf06dc03d2c7acdc9b12f676414ede12cf0e2731d82fb85a3729c46005"} Oct 09 10:15:41 crc kubenswrapper[4923]: I1009 10:15:41.046068 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerStarted","Data":"0e78e72c64d284f1d6dc4bf638b61f057f6ee515caa60ffe857ab8e5aef92f6f"} Oct 09 10:15:41 crc kubenswrapper[4923]: I1009 10:15:41.046084 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerStarted","Data":"fc2a7c716fd6f38b62f7d384e320cd492698350ef3247a0bd5182255620a2259"} Oct 09 10:15:41 crc kubenswrapper[4923]: I1009 10:15:41.046095 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerStarted","Data":"376302c75e51365dfaa8e5ca044f404c92214db506e68e0dfe5081439c204b77"} Oct 09 10:15:41 crc kubenswrapper[4923]: I1009 10:15:41.046105 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerStarted","Data":"209b00f94e90a0e1dadd2534ca4803c7f06cd908b5f72ab590f720dd1ba88bb6"} Oct 09 10:15:43 crc kubenswrapper[4923]: I1009 10:15:43.064292 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerStarted","Data":"12d2e7b5ed144d6dd5d03ab6fb8887bb6c1367888a079853e297c67ca414d884"} Oct 09 10:15:46 crc kubenswrapper[4923]: I1009 10:15:46.089586 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" event={"ID":"f0d84dc9-d157-41c4-a85e-1daba23b31ce","Type":"ContainerStarted","Data":"fa68de10c393c47a847fdb106807505affe766f5c47c04daaae4df313995e9bc"} Oct 09 10:15:46 crc kubenswrapper[4923]: I1009 10:15:46.090330 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:46 crc kubenswrapper[4923]: I1009 10:15:46.090345 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:46 crc kubenswrapper[4923]: I1009 10:15:46.122569 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:46 crc kubenswrapper[4923]: I1009 10:15:46.132023 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" podStartSLOduration=8.13197904 podStartE2EDuration="8.13197904s" podCreationTimestamp="2025-10-09 10:15:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:15:46.125128499 +0000 UTC m=+632.193310275" watchObservedRunningTime="2025-10-09 10:15:46.13197904 +0000 UTC m=+632.200160836" Oct 09 10:15:47 crc kubenswrapper[4923]: I1009 10:15:47.098987 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:47 crc kubenswrapper[4923]: I1009 10:15:47.133151 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:15:53 crc kubenswrapper[4923]: I1009 10:15:53.601500 4923 scope.go:117] "RemoveContainer" containerID="8aa38bf09b127046fbe3053995133ff9c81ee52e4a4cd7e17444e49b1e6ce556" Oct 09 10:15:53 crc kubenswrapper[4923]: E1009 10:15:53.602306 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-4cn4f_openshift-multus(d5c5f455-4e94-4f9a-b6fa-2535685c6354)\"" pod="openshift-multus/multus-4cn4f" podUID="d5c5f455-4e94-4f9a-b6fa-2535685c6354" Oct 09 10:16:06 crc kubenswrapper[4923]: I1009 10:16:06.602197 4923 scope.go:117] "RemoveContainer" containerID="8aa38bf09b127046fbe3053995133ff9c81ee52e4a4cd7e17444e49b1e6ce556" Oct 09 10:16:07 crc kubenswrapper[4923]: I1009 10:16:07.232904 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/2.log" Oct 09 10:16:07 crc kubenswrapper[4923]: I1009 10:16:07.234272 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/1.log" Oct 09 10:16:07 crc kubenswrapper[4923]: I1009 10:16:07.234363 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-4cn4f" event={"ID":"d5c5f455-4e94-4f9a-b6fa-2535685c6354","Type":"ContainerStarted","Data":"c0c9224542249842feed5fe98e0324626cbd46c90117d6dc5c78d9d6dccc27f4"} Oct 09 10:16:08 crc kubenswrapper[4923]: I1009 10:16:08.888333 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-tp98r" Oct 09 10:16:14 crc kubenswrapper[4923]: I1009 10:16:14.803966 4923 scope.go:117] "RemoveContainer" containerID="29124e378155247dbf1f6dc839294ef018fba4e299fee7ea47c17875b396d17a" Oct 09 10:16:15 crc kubenswrapper[4923]: I1009 10:16:15.289464 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-4cn4f_d5c5f455-4e94-4f9a-b6fa-2535685c6354/kube-multus/2.log" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.417307 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg"] Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.419104 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.422774 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.429469 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg"] Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.588829 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2566\" (UniqueName: \"kubernetes.io/projected/4c7a6073-930b-477d-a8af-67196a768b71-kube-api-access-q2566\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.588971 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.589013 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.691185 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2566\" (UniqueName: \"kubernetes.io/projected/4c7a6073-930b-477d-a8af-67196a768b71-kube-api-access-q2566\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.691910 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.692831 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.694589 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.694725 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.717645 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2566\" (UniqueName: \"kubernetes.io/projected/4c7a6073-930b-477d-a8af-67196a768b71-kube-api-access-q2566\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:22 crc kubenswrapper[4923]: I1009 10:16:22.749805 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:23 crc kubenswrapper[4923]: I1009 10:16:23.021263 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg"] Oct 09 10:16:23 crc kubenswrapper[4923]: W1009 10:16:23.029482 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c7a6073_930b_477d_a8af_67196a768b71.slice/crio-47822e17da3a0bbec0b9562ba4e02541369ac1e49449bec29023bd6aa7a647cc WatchSource:0}: Error finding container 47822e17da3a0bbec0b9562ba4e02541369ac1e49449bec29023bd6aa7a647cc: Status 404 returned error can't find the container with id 47822e17da3a0bbec0b9562ba4e02541369ac1e49449bec29023bd6aa7a647cc Oct 09 10:16:23 crc kubenswrapper[4923]: I1009 10:16:23.345056 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" event={"ID":"4c7a6073-930b-477d-a8af-67196a768b71","Type":"ContainerStarted","Data":"b8a586950cd4b3e5be80816b69791eb8ddaa0d8c2dc49bd46ea38b7c309fc16f"} Oct 09 10:16:23 crc kubenswrapper[4923]: I1009 10:16:23.345111 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" event={"ID":"4c7a6073-930b-477d-a8af-67196a768b71","Type":"ContainerStarted","Data":"47822e17da3a0bbec0b9562ba4e02541369ac1e49449bec29023bd6aa7a647cc"} Oct 09 10:16:24 crc kubenswrapper[4923]: I1009 10:16:24.357013 4923 generic.go:334] "Generic (PLEG): container finished" podID="4c7a6073-930b-477d-a8af-67196a768b71" containerID="b8a586950cd4b3e5be80816b69791eb8ddaa0d8c2dc49bd46ea38b7c309fc16f" exitCode=0 Oct 09 10:16:24 crc kubenswrapper[4923]: I1009 10:16:24.357427 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" event={"ID":"4c7a6073-930b-477d-a8af-67196a768b71","Type":"ContainerDied","Data":"b8a586950cd4b3e5be80816b69791eb8ddaa0d8c2dc49bd46ea38b7c309fc16f"} Oct 09 10:16:26 crc kubenswrapper[4923]: I1009 10:16:26.374472 4923 generic.go:334] "Generic (PLEG): container finished" podID="4c7a6073-930b-477d-a8af-67196a768b71" containerID="e8e5f604311c4c5e754cdac568c852e8277ee8aae02160765c199028ab955bdd" exitCode=0 Oct 09 10:16:26 crc kubenswrapper[4923]: I1009 10:16:26.374577 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" event={"ID":"4c7a6073-930b-477d-a8af-67196a768b71","Type":"ContainerDied","Data":"e8e5f604311c4c5e754cdac568c852e8277ee8aae02160765c199028ab955bdd"} Oct 09 10:16:27 crc kubenswrapper[4923]: I1009 10:16:27.384558 4923 generic.go:334] "Generic (PLEG): container finished" podID="4c7a6073-930b-477d-a8af-67196a768b71" containerID="2ad252f388ca3b31ed676f8f2b71e51d7a06b5ed2b9f4e62c1a4ae39ffca017a" exitCode=0 Oct 09 10:16:27 crc kubenswrapper[4923]: I1009 10:16:27.384712 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" event={"ID":"4c7a6073-930b-477d-a8af-67196a768b71","Type":"ContainerDied","Data":"2ad252f388ca3b31ed676f8f2b71e51d7a06b5ed2b9f4e62c1a4ae39ffca017a"} Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.673443 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.684457 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-bundle\") pod \"4c7a6073-930b-477d-a8af-67196a768b71\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.684575 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2566\" (UniqueName: \"kubernetes.io/projected/4c7a6073-930b-477d-a8af-67196a768b71-kube-api-access-q2566\") pod \"4c7a6073-930b-477d-a8af-67196a768b71\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.684594 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-util\") pod \"4c7a6073-930b-477d-a8af-67196a768b71\" (UID: \"4c7a6073-930b-477d-a8af-67196a768b71\") " Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.685607 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-bundle" (OuterVolumeSpecName: "bundle") pod "4c7a6073-930b-477d-a8af-67196a768b71" (UID: "4c7a6073-930b-477d-a8af-67196a768b71"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.695268 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c7a6073-930b-477d-a8af-67196a768b71-kube-api-access-q2566" (OuterVolumeSpecName: "kube-api-access-q2566") pod "4c7a6073-930b-477d-a8af-67196a768b71" (UID: "4c7a6073-930b-477d-a8af-67196a768b71"). InnerVolumeSpecName "kube-api-access-q2566". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.786016 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2566\" (UniqueName: \"kubernetes.io/projected/4c7a6073-930b-477d-a8af-67196a768b71-kube-api-access-q2566\") on node \"crc\" DevicePath \"\"" Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.786059 4923 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.857249 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-util" (OuterVolumeSpecName: "util") pod "4c7a6073-930b-477d-a8af-67196a768b71" (UID: "4c7a6073-930b-477d-a8af-67196a768b71"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:16:28 crc kubenswrapper[4923]: I1009 10:16:28.888273 4923 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/4c7a6073-930b-477d-a8af-67196a768b71-util\") on node \"crc\" DevicePath \"\"" Oct 09 10:16:29 crc kubenswrapper[4923]: I1009 10:16:29.400103 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" event={"ID":"4c7a6073-930b-477d-a8af-67196a768b71","Type":"ContainerDied","Data":"47822e17da3a0bbec0b9562ba4e02541369ac1e49449bec29023bd6aa7a647cc"} Oct 09 10:16:29 crc kubenswrapper[4923]: I1009 10:16:29.400147 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg" Oct 09 10:16:29 crc kubenswrapper[4923]: I1009 10:16:29.400164 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47822e17da3a0bbec0b9562ba4e02541369ac1e49449bec29023bd6aa7a647cc" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.074588 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l"] Oct 09 10:16:31 crc kubenswrapper[4923]: E1009 10:16:31.074821 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7a6073-930b-477d-a8af-67196a768b71" containerName="pull" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.074835 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7a6073-930b-477d-a8af-67196a768b71" containerName="pull" Oct 09 10:16:31 crc kubenswrapper[4923]: E1009 10:16:31.074849 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7a6073-930b-477d-a8af-67196a768b71" containerName="util" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.074858 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7a6073-930b-477d-a8af-67196a768b71" containerName="util" Oct 09 10:16:31 crc kubenswrapper[4923]: E1009 10:16:31.074870 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c7a6073-930b-477d-a8af-67196a768b71" containerName="extract" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.074878 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c7a6073-930b-477d-a8af-67196a768b71" containerName="extract" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.074988 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c7a6073-930b-477d-a8af-67196a768b71" containerName="extract" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.075360 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.077746 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.078250 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.079027 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-6xm72" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.099125 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l"] Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.124895 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw78l\" (UniqueName: \"kubernetes.io/projected/fd8d3d9c-df27-445f-8ba6-5c39074945bf-kube-api-access-hw78l\") pod \"nmstate-operator-858ddd8f98-6sg6l\" (UID: \"fd8d3d9c-df27-445f-8ba6-5c39074945bf\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.226702 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw78l\" (UniqueName: \"kubernetes.io/projected/fd8d3d9c-df27-445f-8ba6-5c39074945bf-kube-api-access-hw78l\") pod \"nmstate-operator-858ddd8f98-6sg6l\" (UID: \"fd8d3d9c-df27-445f-8ba6-5c39074945bf\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.245921 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw78l\" (UniqueName: \"kubernetes.io/projected/fd8d3d9c-df27-445f-8ba6-5c39074945bf-kube-api-access-hw78l\") pod \"nmstate-operator-858ddd8f98-6sg6l\" (UID: \"fd8d3d9c-df27-445f-8ba6-5c39074945bf\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.394619 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l" Oct 09 10:16:31 crc kubenswrapper[4923]: I1009 10:16:31.865258 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l"] Oct 09 10:16:32 crc kubenswrapper[4923]: I1009 10:16:32.425333 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l" event={"ID":"fd8d3d9c-df27-445f-8ba6-5c39074945bf","Type":"ContainerStarted","Data":"176e307de184b3f8d3522c2d6cf978dcecd85ccfecd5069ebd8f77878e6b0f49"} Oct 09 10:16:34 crc kubenswrapper[4923]: I1009 10:16:34.441101 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l" event={"ID":"fd8d3d9c-df27-445f-8ba6-5c39074945bf","Type":"ContainerStarted","Data":"f231d149e68ebb5ef353863c6c8a91cc386f84a83d764ebe2d66b3a66caa8d60"} Oct 09 10:16:34 crc kubenswrapper[4923]: I1009 10:16:34.472632 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-6sg6l" podStartSLOduration=1.115011369 podStartE2EDuration="3.472596617s" podCreationTimestamp="2025-10-09 10:16:31 +0000 UTC" firstStartedPulling="2025-10-09 10:16:31.879779214 +0000 UTC m=+677.947960970" lastFinishedPulling="2025-10-09 10:16:34.237364462 +0000 UTC m=+680.305546218" observedRunningTime="2025-10-09 10:16:34.46985168 +0000 UTC m=+680.538033436" watchObservedRunningTime="2025-10-09 10:16:34.472596617 +0000 UTC m=+680.540778373" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.481931 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l"] Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.483416 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.485860 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-s2zws" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.485988 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzqxv\" (UniqueName: \"kubernetes.io/projected/9170945d-01c4-4c87-a7fa-c5220a39acdd-kube-api-access-mzqxv\") pod \"nmstate-metrics-fdff9cb8d-xsz5l\" (UID: \"9170945d-01c4-4c87-a7fa-c5220a39acdd\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.500910 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l"] Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.508090 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd"] Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.508793 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.512640 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.532262 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-9t6jc"] Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.532956 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.540300 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd"] Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.589071 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8e7041e7-f857-461a-a327-3a267fdfe83b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-r8kkd\" (UID: \"8e7041e7-f857-461a-a327-3a267fdfe83b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.589113 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/59a2d879-dfd9-4342-aef1-d99ebcbc847e-dbus-socket\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.589138 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/59a2d879-dfd9-4342-aef1-d99ebcbc847e-nmstate-lock\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.589160 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/59a2d879-dfd9-4342-aef1-d99ebcbc847e-ovs-socket\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.589195 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpbkx\" (UniqueName: \"kubernetes.io/projected/59a2d879-dfd9-4342-aef1-d99ebcbc847e-kube-api-access-xpbkx\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.589223 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fdrl\" (UniqueName: \"kubernetes.io/projected/8e7041e7-f857-461a-a327-3a267fdfe83b-kube-api-access-4fdrl\") pod \"nmstate-webhook-6cdbc54649-r8kkd\" (UID: \"8e7041e7-f857-461a-a327-3a267fdfe83b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.589259 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzqxv\" (UniqueName: \"kubernetes.io/projected/9170945d-01c4-4c87-a7fa-c5220a39acdd-kube-api-access-mzqxv\") pod \"nmstate-metrics-fdff9cb8d-xsz5l\" (UID: \"9170945d-01c4-4c87-a7fa-c5220a39acdd\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.635422 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzqxv\" (UniqueName: \"kubernetes.io/projected/9170945d-01c4-4c87-a7fa-c5220a39acdd-kube-api-access-mzqxv\") pod \"nmstate-metrics-fdff9cb8d-xsz5l\" (UID: \"9170945d-01c4-4c87-a7fa-c5220a39acdd\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.665291 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6"] Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.666381 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.669558 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-sql4n" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.670299 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.671845 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.684856 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6"] Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.690231 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/685e4264-fd07-4f0c-8223-7f4ac22e89b8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.690291 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpbkx\" (UniqueName: \"kubernetes.io/projected/59a2d879-dfd9-4342-aef1-d99ebcbc847e-kube-api-access-xpbkx\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.690492 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dpbm\" (UniqueName: \"kubernetes.io/projected/685e4264-fd07-4f0c-8223-7f4ac22e89b8-kube-api-access-2dpbm\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.690685 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fdrl\" (UniqueName: \"kubernetes.io/projected/8e7041e7-f857-461a-a327-3a267fdfe83b-kube-api-access-4fdrl\") pod \"nmstate-webhook-6cdbc54649-r8kkd\" (UID: \"8e7041e7-f857-461a-a327-3a267fdfe83b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.690820 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/685e4264-fd07-4f0c-8223-7f4ac22e89b8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.690966 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8e7041e7-f857-461a-a327-3a267fdfe83b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-r8kkd\" (UID: \"8e7041e7-f857-461a-a327-3a267fdfe83b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.691070 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/59a2d879-dfd9-4342-aef1-d99ebcbc847e-dbus-socket\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.691191 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/59a2d879-dfd9-4342-aef1-d99ebcbc847e-nmstate-lock\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.691315 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/59a2d879-dfd9-4342-aef1-d99ebcbc847e-ovs-socket\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.691468 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/59a2d879-dfd9-4342-aef1-d99ebcbc847e-ovs-socket\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.691585 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/59a2d879-dfd9-4342-aef1-d99ebcbc847e-nmstate-lock\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: E1009 10:16:35.691747 4923 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 09 10:16:35 crc kubenswrapper[4923]: E1009 10:16:35.691891 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e7041e7-f857-461a-a327-3a267fdfe83b-tls-key-pair podName:8e7041e7-f857-461a-a327-3a267fdfe83b nodeName:}" failed. No retries permitted until 2025-10-09 10:16:36.191868753 +0000 UTC m=+682.260050989 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/8e7041e7-f857-461a-a327-3a267fdfe83b-tls-key-pair") pod "nmstate-webhook-6cdbc54649-r8kkd" (UID: "8e7041e7-f857-461a-a327-3a267fdfe83b") : secret "openshift-nmstate-webhook" not found Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.692733 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/59a2d879-dfd9-4342-aef1-d99ebcbc847e-dbus-socket\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.714670 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpbkx\" (UniqueName: \"kubernetes.io/projected/59a2d879-dfd9-4342-aef1-d99ebcbc847e-kube-api-access-xpbkx\") pod \"nmstate-handler-9t6jc\" (UID: \"59a2d879-dfd9-4342-aef1-d99ebcbc847e\") " pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.720745 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fdrl\" (UniqueName: \"kubernetes.io/projected/8e7041e7-f857-461a-a327-3a267fdfe83b-kube-api-access-4fdrl\") pod \"nmstate-webhook-6cdbc54649-r8kkd\" (UID: \"8e7041e7-f857-461a-a327-3a267fdfe83b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.791933 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/685e4264-fd07-4f0c-8223-7f4ac22e89b8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.791996 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dpbm\" (UniqueName: \"kubernetes.io/projected/685e4264-fd07-4f0c-8223-7f4ac22e89b8-kube-api-access-2dpbm\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.792022 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/685e4264-fd07-4f0c-8223-7f4ac22e89b8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:35 crc kubenswrapper[4923]: E1009 10:16:35.792222 4923 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Oct 09 10:16:35 crc kubenswrapper[4923]: E1009 10:16:35.792306 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/685e4264-fd07-4f0c-8223-7f4ac22e89b8-plugin-serving-cert podName:685e4264-fd07-4f0c-8223-7f4ac22e89b8 nodeName:}" failed. No retries permitted until 2025-10-09 10:16:36.292285332 +0000 UTC m=+682.360467088 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/685e4264-fd07-4f0c-8223-7f4ac22e89b8-plugin-serving-cert") pod "nmstate-console-plugin-6b874cbd85-5swk6" (UID: "685e4264-fd07-4f0c-8223-7f4ac22e89b8") : secret "plugin-serving-cert" not found Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.793252 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/685e4264-fd07-4f0c-8223-7f4ac22e89b8-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.802236 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.822428 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dpbm\" (UniqueName: \"kubernetes.io/projected/685e4264-fd07-4f0c-8223-7f4ac22e89b8-kube-api-access-2dpbm\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.851350 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.888943 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6cc9ff59d7-l5tqm"] Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.889788 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.895563 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-console-serving-cert\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.895620 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-console-config\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.895677 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpf27\" (UniqueName: \"kubernetes.io/projected/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-kube-api-access-zpf27\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.895758 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-console-oauth-config\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.895822 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-service-ca\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.895860 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-oauth-serving-cert\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.895886 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-trusted-ca-bundle\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.912412 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6cc9ff59d7-l5tqm"] Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.997707 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-trusted-ca-bundle\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.997795 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-console-serving-cert\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.997830 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-console-config\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.997885 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpf27\" (UniqueName: \"kubernetes.io/projected/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-kube-api-access-zpf27\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.997926 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-console-oauth-config\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.997955 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-service-ca\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:35 crc kubenswrapper[4923]: I1009 10:16:35.997998 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-oauth-serving-cert\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.000647 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-oauth-serving-cert\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.004168 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-trusted-ca-bundle\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.004459 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-console-oauth-config\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.005521 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-service-ca\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.005829 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-console-serving-cert\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.006273 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-console-config\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.034504 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpf27\" (UniqueName: \"kubernetes.io/projected/a22f0a80-9fd7-4832-acc3-a790f1f31cc5-kube-api-access-zpf27\") pod \"console-6cc9ff59d7-l5tqm\" (UID: \"a22f0a80-9fd7-4832-acc3-a790f1f31cc5\") " pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.090321 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l"] Oct 09 10:16:36 crc kubenswrapper[4923]: W1009 10:16:36.098123 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9170945d_01c4_4c87_a7fa_c5220a39acdd.slice/crio-6cd953e92994acc704210fe79d65c450468d45c0e1cab84419fa74649863e554 WatchSource:0}: Error finding container 6cd953e92994acc704210fe79d65c450468d45c0e1cab84419fa74649863e554: Status 404 returned error can't find the container with id 6cd953e92994acc704210fe79d65c450468d45c0e1cab84419fa74649863e554 Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.199973 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8e7041e7-f857-461a-a327-3a267fdfe83b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-r8kkd\" (UID: \"8e7041e7-f857-461a-a327-3a267fdfe83b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.205173 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8e7041e7-f857-461a-a327-3a267fdfe83b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-r8kkd\" (UID: \"8e7041e7-f857-461a-a327-3a267fdfe83b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.215648 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.301949 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/685e4264-fd07-4f0c-8223-7f4ac22e89b8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.306261 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/685e4264-fd07-4f0c-8223-7f4ac22e89b8-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-5swk6\" (UID: \"685e4264-fd07-4f0c-8223-7f4ac22e89b8\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.415992 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6cc9ff59d7-l5tqm"] Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.425638 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:36 crc kubenswrapper[4923]: W1009 10:16:36.425832 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda22f0a80_9fd7_4832_acc3_a790f1f31cc5.slice/crio-92fb2478d0b2e6b21099fee6740b27d59800ed676f8699f6d020009ed72d45c0 WatchSource:0}: Error finding container 92fb2478d0b2e6b21099fee6740b27d59800ed676f8699f6d020009ed72d45c0: Status 404 returned error can't find the container with id 92fb2478d0b2e6b21099fee6740b27d59800ed676f8699f6d020009ed72d45c0 Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.457120 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6cc9ff59d7-l5tqm" event={"ID":"a22f0a80-9fd7-4832-acc3-a790f1f31cc5","Type":"ContainerStarted","Data":"92fb2478d0b2e6b21099fee6740b27d59800ed676f8699f6d020009ed72d45c0"} Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.458204 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-9t6jc" event={"ID":"59a2d879-dfd9-4342-aef1-d99ebcbc847e","Type":"ContainerStarted","Data":"1862188d0742dbc8a51887f6481edfc57de3deee64ad8f5fea4d392661f20f5e"} Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.459429 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l" event={"ID":"9170945d-01c4-4c87-a7fa-c5220a39acdd","Type":"ContainerStarted","Data":"6cd953e92994acc704210fe79d65c450468d45c0e1cab84419fa74649863e554"} Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.586283 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.635873 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd"] Oct 09 10:16:36 crc kubenswrapper[4923]: W1009 10:16:36.653339 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e7041e7_f857_461a_a327_3a267fdfe83b.slice/crio-804b7d4e6b5ab818bc9483ca3586a59c448fb22f9302b8d8ce6c259e7db6281e WatchSource:0}: Error finding container 804b7d4e6b5ab818bc9483ca3586a59c448fb22f9302b8d8ce6c259e7db6281e: Status 404 returned error can't find the container with id 804b7d4e6b5ab818bc9483ca3586a59c448fb22f9302b8d8ce6c259e7db6281e Oct 09 10:16:36 crc kubenswrapper[4923]: I1009 10:16:36.834980 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6"] Oct 09 10:16:37 crc kubenswrapper[4923]: I1009 10:16:37.468291 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" event={"ID":"8e7041e7-f857-461a-a327-3a267fdfe83b","Type":"ContainerStarted","Data":"804b7d4e6b5ab818bc9483ca3586a59c448fb22f9302b8d8ce6c259e7db6281e"} Oct 09 10:16:37 crc kubenswrapper[4923]: I1009 10:16:37.469420 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" event={"ID":"685e4264-fd07-4f0c-8223-7f4ac22e89b8","Type":"ContainerStarted","Data":"22de793ea436c5002524ffd5a5e3b79cb4cb3298405584d3ffe4ba652388bddb"} Oct 09 10:16:37 crc kubenswrapper[4923]: I1009 10:16:37.470643 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6cc9ff59d7-l5tqm" event={"ID":"a22f0a80-9fd7-4832-acc3-a790f1f31cc5","Type":"ContainerStarted","Data":"f66800c28eb88564ccdb6d18dcbe361f48859e5d1c66aaf56557cfc4c19d01d5"} Oct 09 10:16:37 crc kubenswrapper[4923]: I1009 10:16:37.497795 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6cc9ff59d7-l5tqm" podStartSLOduration=2.497755047 podStartE2EDuration="2.497755047s" podCreationTimestamp="2025-10-09 10:16:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:16:37.491510823 +0000 UTC m=+683.559692579" watchObservedRunningTime="2025-10-09 10:16:37.497755047 +0000 UTC m=+683.565936803" Oct 09 10:16:42 crc kubenswrapper[4923]: I1009 10:16:42.529949 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" event={"ID":"8e7041e7-f857-461a-a327-3a267fdfe83b","Type":"ContainerStarted","Data":"245e5398414cda62a8e3be45e5036e105bf20912019bdc5d93e728709c14e7c2"} Oct 09 10:16:42 crc kubenswrapper[4923]: I1009 10:16:42.530809 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:16:42 crc kubenswrapper[4923]: I1009 10:16:42.532900 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" event={"ID":"685e4264-fd07-4f0c-8223-7f4ac22e89b8","Type":"ContainerStarted","Data":"135d698faae6bf54ece8e3a8d1e5580397cf61b148c8685cbe4c672fbd946181"} Oct 09 10:16:42 crc kubenswrapper[4923]: I1009 10:16:42.536743 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-9t6jc" event={"ID":"59a2d879-dfd9-4342-aef1-d99ebcbc847e","Type":"ContainerStarted","Data":"1e2b29b60699e81bcfe61aabe6f1fa052f6b094eb7f0274891994f5582249500"} Oct 09 10:16:42 crc kubenswrapper[4923]: I1009 10:16:42.536972 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:42 crc kubenswrapper[4923]: I1009 10:16:42.538011 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l" event={"ID":"9170945d-01c4-4c87-a7fa-c5220a39acdd","Type":"ContainerStarted","Data":"31e5663eb163b0a93bd0e12842d749631b075b1bd8dea8b484ec745b2c7d19c6"} Oct 09 10:16:42 crc kubenswrapper[4923]: I1009 10:16:42.553968 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" podStartSLOduration=2.717548682 podStartE2EDuration="7.553944573s" podCreationTimestamp="2025-10-09 10:16:35 +0000 UTC" firstStartedPulling="2025-10-09 10:16:36.658722536 +0000 UTC m=+682.726904292" lastFinishedPulling="2025-10-09 10:16:41.495118427 +0000 UTC m=+687.563300183" observedRunningTime="2025-10-09 10:16:42.549131608 +0000 UTC m=+688.617313384" watchObservedRunningTime="2025-10-09 10:16:42.553944573 +0000 UTC m=+688.622126329" Oct 09 10:16:42 crc kubenswrapper[4923]: I1009 10:16:42.583066 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-5swk6" podStartSLOduration=2.9705551 podStartE2EDuration="7.583046223s" podCreationTimestamp="2025-10-09 10:16:35 +0000 UTC" firstStartedPulling="2025-10-09 10:16:36.846285102 +0000 UTC m=+682.914466848" lastFinishedPulling="2025-10-09 10:16:41.458776205 +0000 UTC m=+687.526957971" observedRunningTime="2025-10-09 10:16:42.579397892 +0000 UTC m=+688.647579648" watchObservedRunningTime="2025-10-09 10:16:42.583046223 +0000 UTC m=+688.651227979" Oct 09 10:16:42 crc kubenswrapper[4923]: I1009 10:16:42.613951 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-9t6jc" podStartSLOduration=2.047460198 podStartE2EDuration="7.613930174s" podCreationTimestamp="2025-10-09 10:16:35 +0000 UTC" firstStartedPulling="2025-10-09 10:16:35.893899533 +0000 UTC m=+681.962081289" lastFinishedPulling="2025-10-09 10:16:41.460369469 +0000 UTC m=+687.528551265" observedRunningTime="2025-10-09 10:16:42.596695774 +0000 UTC m=+688.664877530" watchObservedRunningTime="2025-10-09 10:16:42.613930174 +0000 UTC m=+688.682111920" Oct 09 10:16:45 crc kubenswrapper[4923]: I1009 10:16:45.561982 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l" event={"ID":"9170945d-01c4-4c87-a7fa-c5220a39acdd","Type":"ContainerStarted","Data":"51459f4585f62f4bb71c88e07be565a26e580c99a011aa24a3492bc7a97a9d58"} Oct 09 10:16:45 crc kubenswrapper[4923]: I1009 10:16:45.586824 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-xsz5l" podStartSLOduration=1.3344005669999999 podStartE2EDuration="10.586800566s" podCreationTimestamp="2025-10-09 10:16:35 +0000 UTC" firstStartedPulling="2025-10-09 10:16:36.100411148 +0000 UTC m=+682.168592904" lastFinishedPulling="2025-10-09 10:16:45.352811137 +0000 UTC m=+691.420992903" observedRunningTime="2025-10-09 10:16:45.580718377 +0000 UTC m=+691.648900173" watchObservedRunningTime="2025-10-09 10:16:45.586800566 +0000 UTC m=+691.654982312" Oct 09 10:16:46 crc kubenswrapper[4923]: I1009 10:16:46.216382 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:46 crc kubenswrapper[4923]: I1009 10:16:46.217127 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:46 crc kubenswrapper[4923]: I1009 10:16:46.242331 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:46 crc kubenswrapper[4923]: I1009 10:16:46.581317 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6cc9ff59d7-l5tqm" Oct 09 10:16:46 crc kubenswrapper[4923]: I1009 10:16:46.653619 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-66f2z"] Oct 09 10:16:50 crc kubenswrapper[4923]: I1009 10:16:50.887722 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-9t6jc" Oct 09 10:16:56 crc kubenswrapper[4923]: I1009 10:16:56.435716 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-r8kkd" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.674437 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn"] Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.676464 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.678850 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.688708 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn"] Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.714923 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gm5nq\" (UniqueName: \"kubernetes.io/projected/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-kube-api-access-gm5nq\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.715015 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.715062 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.816759 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.816843 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.816914 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gm5nq\" (UniqueName: \"kubernetes.io/projected/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-kube-api-access-gm5nq\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.817307 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.817368 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.839108 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gm5nq\" (UniqueName: \"kubernetes.io/projected/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-kube-api-access-gm5nq\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:10 crc kubenswrapper[4923]: I1009 10:17:10.994701 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:11 crc kubenswrapper[4923]: I1009 10:17:11.395439 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn"] Oct 09 10:17:11 crc kubenswrapper[4923]: I1009 10:17:11.708828 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-66f2z" podUID="a8c487fa-3436-4cc9-83b3-de49256e97c9" containerName="console" containerID="cri-o://564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6" gracePeriod=15 Oct 09 10:17:11 crc kubenswrapper[4923]: I1009 10:17:11.747869 4923 generic.go:334] "Generic (PLEG): container finished" podID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerID="1cc0670503ecdcf2adc740fa8a2111033f483fdd6d2e07141a45db9b19b45a4c" exitCode=0 Oct 09 10:17:11 crc kubenswrapper[4923]: I1009 10:17:11.747920 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" event={"ID":"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439","Type":"ContainerDied","Data":"1cc0670503ecdcf2adc740fa8a2111033f483fdd6d2e07141a45db9b19b45a4c"} Oct 09 10:17:11 crc kubenswrapper[4923]: I1009 10:17:11.747951 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" event={"ID":"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439","Type":"ContainerStarted","Data":"0b0e4ebb8bc89a3a96026eea32586515c89731b47e3374aaae38b44a7612a641"} Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.093410 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-66f2z_a8c487fa-3436-4cc9-83b3-de49256e97c9/console/0.log" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.093855 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.235994 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-oauth-config\") pod \"a8c487fa-3436-4cc9-83b3-de49256e97c9\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.236094 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bh754\" (UniqueName: \"kubernetes.io/projected/a8c487fa-3436-4cc9-83b3-de49256e97c9-kube-api-access-bh754\") pod \"a8c487fa-3436-4cc9-83b3-de49256e97c9\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.236142 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-trusted-ca-bundle\") pod \"a8c487fa-3436-4cc9-83b3-de49256e97c9\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.236161 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-config\") pod \"a8c487fa-3436-4cc9-83b3-de49256e97c9\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.236193 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-oauth-serving-cert\") pod \"a8c487fa-3436-4cc9-83b3-de49256e97c9\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.236221 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-serving-cert\") pod \"a8c487fa-3436-4cc9-83b3-de49256e97c9\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.236259 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-service-ca\") pod \"a8c487fa-3436-4cc9-83b3-de49256e97c9\" (UID: \"a8c487fa-3436-4cc9-83b3-de49256e97c9\") " Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.237374 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a8c487fa-3436-4cc9-83b3-de49256e97c9" (UID: "a8c487fa-3436-4cc9-83b3-de49256e97c9"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.237600 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-config" (OuterVolumeSpecName: "console-config") pod "a8c487fa-3436-4cc9-83b3-de49256e97c9" (UID: "a8c487fa-3436-4cc9-83b3-de49256e97c9"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.237626 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-service-ca" (OuterVolumeSpecName: "service-ca") pod "a8c487fa-3436-4cc9-83b3-de49256e97c9" (UID: "a8c487fa-3436-4cc9-83b3-de49256e97c9"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.237795 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a8c487fa-3436-4cc9-83b3-de49256e97c9" (UID: "a8c487fa-3436-4cc9-83b3-de49256e97c9"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.243088 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a8c487fa-3436-4cc9-83b3-de49256e97c9" (UID: "a8c487fa-3436-4cc9-83b3-de49256e97c9"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.243140 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8c487fa-3436-4cc9-83b3-de49256e97c9-kube-api-access-bh754" (OuterVolumeSpecName: "kube-api-access-bh754") pod "a8c487fa-3436-4cc9-83b3-de49256e97c9" (UID: "a8c487fa-3436-4cc9-83b3-de49256e97c9"). InnerVolumeSpecName "kube-api-access-bh754". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.243698 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a8c487fa-3436-4cc9-83b3-de49256e97c9" (UID: "a8c487fa-3436-4cc9-83b3-de49256e97c9"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.338253 4923 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.338310 4923 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.338327 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bh754\" (UniqueName: \"kubernetes.io/projected/a8c487fa-3436-4cc9-83b3-de49256e97c9-kube-api-access-bh754\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.338337 4923 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.338351 4923 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.338359 4923 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a8c487fa-3436-4cc9-83b3-de49256e97c9-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.338369 4923 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a8c487fa-3436-4cc9-83b3-de49256e97c9-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.763416 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-66f2z_a8c487fa-3436-4cc9-83b3-de49256e97c9/console/0.log" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.763501 4923 generic.go:334] "Generic (PLEG): container finished" podID="a8c487fa-3436-4cc9-83b3-de49256e97c9" containerID="564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6" exitCode=2 Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.763587 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-66f2z" event={"ID":"a8c487fa-3436-4cc9-83b3-de49256e97c9","Type":"ContainerDied","Data":"564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6"} Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.763648 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-66f2z" event={"ID":"a8c487fa-3436-4cc9-83b3-de49256e97c9","Type":"ContainerDied","Data":"cb8e7efd2a6f964b7d205e396df38dce86f7f139beb9137f97b52b007f3113b8"} Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.763671 4923 scope.go:117] "RemoveContainer" containerID="564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.763905 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-66f2z" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.782870 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-66f2z"] Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.786059 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-66f2z"] Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.832918 4923 scope.go:117] "RemoveContainer" containerID="564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6" Oct 09 10:17:12 crc kubenswrapper[4923]: E1009 10:17:12.834191 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6\": container with ID starting with 564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6 not found: ID does not exist" containerID="564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6" Oct 09 10:17:12 crc kubenswrapper[4923]: I1009 10:17:12.834253 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6"} err="failed to get container status \"564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6\": rpc error: code = NotFound desc = could not find container \"564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6\": container with ID starting with 564c77c484f3c4f09f339c8f2d16360dfc4fc8e0bb9d22d954ac360f35b3c9f6 not found: ID does not exist" Oct 09 10:17:13 crc kubenswrapper[4923]: I1009 10:17:13.776230 4923 generic.go:334] "Generic (PLEG): container finished" podID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerID="71368370fe99da8425885c320c8630d641b6e3f3cbeabdc299c927ee81ff949b" exitCode=0 Oct 09 10:17:13 crc kubenswrapper[4923]: I1009 10:17:13.776286 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" event={"ID":"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439","Type":"ContainerDied","Data":"71368370fe99da8425885c320c8630d641b6e3f3cbeabdc299c927ee81ff949b"} Oct 09 10:17:14 crc kubenswrapper[4923]: I1009 10:17:14.612118 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8c487fa-3436-4cc9-83b3-de49256e97c9" path="/var/lib/kubelet/pods/a8c487fa-3436-4cc9-83b3-de49256e97c9/volumes" Oct 09 10:17:14 crc kubenswrapper[4923]: I1009 10:17:14.784653 4923 generic.go:334] "Generic (PLEG): container finished" podID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerID="8ca05c0e26b03dbaa8d59f2a0fe2f0fca26759d2be6dcdaffcda90567640526d" exitCode=0 Oct 09 10:17:14 crc kubenswrapper[4923]: I1009 10:17:14.784712 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" event={"ID":"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439","Type":"ContainerDied","Data":"8ca05c0e26b03dbaa8d59f2a0fe2f0fca26759d2be6dcdaffcda90567640526d"} Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.018953 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.192548 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gm5nq\" (UniqueName: \"kubernetes.io/projected/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-kube-api-access-gm5nq\") pod \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.192615 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-util\") pod \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.192730 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-bundle\") pod \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\" (UID: \"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439\") " Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.194289 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-bundle" (OuterVolumeSpecName: "bundle") pod "7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" (UID: "7c8cda6e-95c6-4907-8aa0-6bbf50a0c439"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.200892 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-kube-api-access-gm5nq" (OuterVolumeSpecName: "kube-api-access-gm5nq") pod "7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" (UID: "7c8cda6e-95c6-4907-8aa0-6bbf50a0c439"). InnerVolumeSpecName "kube-api-access-gm5nq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.206944 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-util" (OuterVolumeSpecName: "util") pod "7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" (UID: "7c8cda6e-95c6-4907-8aa0-6bbf50a0c439"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.294083 4923 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.294133 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gm5nq\" (UniqueName: \"kubernetes.io/projected/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-kube-api-access-gm5nq\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.294152 4923 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7c8cda6e-95c6-4907-8aa0-6bbf50a0c439-util\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.800985 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.801845 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn" event={"ID":"7c8cda6e-95c6-4907-8aa0-6bbf50a0c439","Type":"ContainerDied","Data":"0b0e4ebb8bc89a3a96026eea32586515c89731b47e3374aaae38b44a7612a641"} Oct 09 10:17:16 crc kubenswrapper[4923]: I1009 10:17:16.801917 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b0e4ebb8bc89a3a96026eea32586515c89731b47e3374aaae38b44a7612a641" Oct 09 10:17:24 crc kubenswrapper[4923]: I1009 10:17:24.600190 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:17:24 crc kubenswrapper[4923]: I1009 10:17:24.601092 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.227704 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5"] Oct 09 10:17:25 crc kubenswrapper[4923]: E1009 10:17:25.228019 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerName="util" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.228036 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerName="util" Oct 09 10:17:25 crc kubenswrapper[4923]: E1009 10:17:25.228056 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerName="pull" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.228064 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerName="pull" Oct 09 10:17:25 crc kubenswrapper[4923]: E1009 10:17:25.228076 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c487fa-3436-4cc9-83b3-de49256e97c9" containerName="console" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.228084 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c487fa-3436-4cc9-83b3-de49256e97c9" containerName="console" Oct 09 10:17:25 crc kubenswrapper[4923]: E1009 10:17:25.228100 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerName="extract" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.228108 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerName="extract" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.228231 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c8cda6e-95c6-4907-8aa0-6bbf50a0c439" containerName="extract" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.228246 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8c487fa-3436-4cc9-83b3-de49256e97c9" containerName="console" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.228787 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.230784 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.230803 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.231272 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.234137 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.234720 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-trzct" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.242605 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5"] Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.324007 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2da4daff-10af-46a7-9b37-dde79cb66313-webhook-cert\") pod \"metallb-operator-controller-manager-59d99554d9-d78r5\" (UID: \"2da4daff-10af-46a7-9b37-dde79cb66313\") " pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.324643 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2da4daff-10af-46a7-9b37-dde79cb66313-apiservice-cert\") pod \"metallb-operator-controller-manager-59d99554d9-d78r5\" (UID: \"2da4daff-10af-46a7-9b37-dde79cb66313\") " pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.324807 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbclg\" (UniqueName: \"kubernetes.io/projected/2da4daff-10af-46a7-9b37-dde79cb66313-kube-api-access-lbclg\") pod \"metallb-operator-controller-manager-59d99554d9-d78r5\" (UID: \"2da4daff-10af-46a7-9b37-dde79cb66313\") " pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.426931 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2da4daff-10af-46a7-9b37-dde79cb66313-webhook-cert\") pod \"metallb-operator-controller-manager-59d99554d9-d78r5\" (UID: \"2da4daff-10af-46a7-9b37-dde79cb66313\") " pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.427007 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2da4daff-10af-46a7-9b37-dde79cb66313-apiservice-cert\") pod \"metallb-operator-controller-manager-59d99554d9-d78r5\" (UID: \"2da4daff-10af-46a7-9b37-dde79cb66313\") " pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.427104 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbclg\" (UniqueName: \"kubernetes.io/projected/2da4daff-10af-46a7-9b37-dde79cb66313-kube-api-access-lbclg\") pod \"metallb-operator-controller-manager-59d99554d9-d78r5\" (UID: \"2da4daff-10af-46a7-9b37-dde79cb66313\") " pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.434549 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2da4daff-10af-46a7-9b37-dde79cb66313-webhook-cert\") pod \"metallb-operator-controller-manager-59d99554d9-d78r5\" (UID: \"2da4daff-10af-46a7-9b37-dde79cb66313\") " pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.442485 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2da4daff-10af-46a7-9b37-dde79cb66313-apiservice-cert\") pod \"metallb-operator-controller-manager-59d99554d9-d78r5\" (UID: \"2da4daff-10af-46a7-9b37-dde79cb66313\") " pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.446562 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbclg\" (UniqueName: \"kubernetes.io/projected/2da4daff-10af-46a7-9b37-dde79cb66313-kube-api-access-lbclg\") pod \"metallb-operator-controller-manager-59d99554d9-d78r5\" (UID: \"2da4daff-10af-46a7-9b37-dde79cb66313\") " pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.479052 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz"] Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.480212 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.485158 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-sdkmx" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.486394 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.486715 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.507418 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz"] Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.529251 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/711d6835-1092-41a2-9de3-476e55baf25c-webhook-cert\") pod \"metallb-operator-webhook-server-b6c67695c-wqcnz\" (UID: \"711d6835-1092-41a2-9de3-476e55baf25c\") " pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.529340 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/711d6835-1092-41a2-9de3-476e55baf25c-apiservice-cert\") pod \"metallb-operator-webhook-server-b6c67695c-wqcnz\" (UID: \"711d6835-1092-41a2-9de3-476e55baf25c\") " pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.529362 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sldrh\" (UniqueName: \"kubernetes.io/projected/711d6835-1092-41a2-9de3-476e55baf25c-kube-api-access-sldrh\") pod \"metallb-operator-webhook-server-b6c67695c-wqcnz\" (UID: \"711d6835-1092-41a2-9de3-476e55baf25c\") " pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.601443 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.630610 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/711d6835-1092-41a2-9de3-476e55baf25c-webhook-cert\") pod \"metallb-operator-webhook-server-b6c67695c-wqcnz\" (UID: \"711d6835-1092-41a2-9de3-476e55baf25c\") " pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.630805 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/711d6835-1092-41a2-9de3-476e55baf25c-apiservice-cert\") pod \"metallb-operator-webhook-server-b6c67695c-wqcnz\" (UID: \"711d6835-1092-41a2-9de3-476e55baf25c\") " pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.630843 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sldrh\" (UniqueName: \"kubernetes.io/projected/711d6835-1092-41a2-9de3-476e55baf25c-kube-api-access-sldrh\") pod \"metallb-operator-webhook-server-b6c67695c-wqcnz\" (UID: \"711d6835-1092-41a2-9de3-476e55baf25c\") " pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.636077 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/711d6835-1092-41a2-9de3-476e55baf25c-webhook-cert\") pod \"metallb-operator-webhook-server-b6c67695c-wqcnz\" (UID: \"711d6835-1092-41a2-9de3-476e55baf25c\") " pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.647549 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/711d6835-1092-41a2-9de3-476e55baf25c-apiservice-cert\") pod \"metallb-operator-webhook-server-b6c67695c-wqcnz\" (UID: \"711d6835-1092-41a2-9de3-476e55baf25c\") " pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.660468 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sldrh\" (UniqueName: \"kubernetes.io/projected/711d6835-1092-41a2-9de3-476e55baf25c-kube-api-access-sldrh\") pod \"metallb-operator-webhook-server-b6c67695c-wqcnz\" (UID: \"711d6835-1092-41a2-9de3-476e55baf25c\") " pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:25 crc kubenswrapper[4923]: I1009 10:17:25.795343 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:26 crc kubenswrapper[4923]: I1009 10:17:26.165455 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5"] Oct 09 10:17:26 crc kubenswrapper[4923]: I1009 10:17:26.169200 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz"] Oct 09 10:17:26 crc kubenswrapper[4923]: W1009 10:17:26.174039 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod711d6835_1092_41a2_9de3_476e55baf25c.slice/crio-a607dba2887b252d081e778e10acba848c5536eccd4441eaeb98a2286b0d1199 WatchSource:0}: Error finding container a607dba2887b252d081e778e10acba848c5536eccd4441eaeb98a2286b0d1199: Status 404 returned error can't find the container with id a607dba2887b252d081e778e10acba848c5536eccd4441eaeb98a2286b0d1199 Oct 09 10:17:26 crc kubenswrapper[4923]: I1009 10:17:26.886707 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" event={"ID":"2da4daff-10af-46a7-9b37-dde79cb66313","Type":"ContainerStarted","Data":"63666eb24fccb6d4df00e780853f449989c8dcf147fad864cb0fecf4ef0bf3e2"} Oct 09 10:17:26 crc kubenswrapper[4923]: I1009 10:17:26.888584 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" event={"ID":"711d6835-1092-41a2-9de3-476e55baf25c","Type":"ContainerStarted","Data":"a607dba2887b252d081e778e10acba848c5536eccd4441eaeb98a2286b0d1199"} Oct 09 10:17:31 crc kubenswrapper[4923]: I1009 10:17:31.932112 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" event={"ID":"2da4daff-10af-46a7-9b37-dde79cb66313","Type":"ContainerStarted","Data":"39989072ff067162a70377dbe97362cf622da0224ee81752e84897e797f7760d"} Oct 09 10:17:31 crc kubenswrapper[4923]: I1009 10:17:31.933227 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:17:31 crc kubenswrapper[4923]: I1009 10:17:31.937156 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" event={"ID":"711d6835-1092-41a2-9de3-476e55baf25c","Type":"ContainerStarted","Data":"cd548cdc0e29fd9a2f1579cb60aefa4e6e539cdbb0be41eea369441fb7a94882"} Oct 09 10:17:31 crc kubenswrapper[4923]: I1009 10:17:31.937807 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:31 crc kubenswrapper[4923]: I1009 10:17:31.959425 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" podStartSLOduration=2.040557352 podStartE2EDuration="6.959402454s" podCreationTimestamp="2025-10-09 10:17:25 +0000 UTC" firstStartedPulling="2025-10-09 10:17:26.16998086 +0000 UTC m=+732.238162616" lastFinishedPulling="2025-10-09 10:17:31.088825962 +0000 UTC m=+737.157007718" observedRunningTime="2025-10-09 10:17:31.955554586 +0000 UTC m=+738.023736342" watchObservedRunningTime="2025-10-09 10:17:31.959402454 +0000 UTC m=+738.027584210" Oct 09 10:17:31 crc kubenswrapper[4923]: I1009 10:17:31.983204 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" podStartSLOduration=2.05155909 podStartE2EDuration="6.983180749s" podCreationTimestamp="2025-10-09 10:17:25 +0000 UTC" firstStartedPulling="2025-10-09 10:17:26.178223311 +0000 UTC m=+732.246405067" lastFinishedPulling="2025-10-09 10:17:31.10984496 +0000 UTC m=+737.178026726" observedRunningTime="2025-10-09 10:17:31.978512259 +0000 UTC m=+738.046694015" watchObservedRunningTime="2025-10-09 10:17:31.983180749 +0000 UTC m=+738.051362495" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.060771 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tnbft"] Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.061892 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" podUID="7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" containerName="controller-manager" containerID="cri-o://0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f" gracePeriod=30 Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.112797 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64"] Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.113090 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" podUID="45725226-1710-4204-8055-ecea69f828ee" containerName="route-controller-manager" containerID="cri-o://9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e" gracePeriod=30 Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.524088 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.528565 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.634489 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-serving-cert\") pod \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.634543 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-config\") pod \"45725226-1710-4204-8055-ecea69f828ee\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.634568 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-client-ca\") pod \"45725226-1710-4204-8055-ecea69f828ee\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.634620 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-client-ca\") pod \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.634649 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-proxy-ca-bundles\") pod \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.634667 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-config\") pod \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.634683 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45725226-1710-4204-8055-ecea69f828ee-serving-cert\") pod \"45725226-1710-4204-8055-ecea69f828ee\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.634715 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nx26v\" (UniqueName: \"kubernetes.io/projected/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-kube-api-access-nx26v\") pod \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\" (UID: \"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44\") " Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.634734 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ssrd\" (UniqueName: \"kubernetes.io/projected/45725226-1710-4204-8055-ecea69f828ee-kube-api-access-4ssrd\") pod \"45725226-1710-4204-8055-ecea69f828ee\" (UID: \"45725226-1710-4204-8055-ecea69f828ee\") " Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.635682 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-client-ca" (OuterVolumeSpecName: "client-ca") pod "45725226-1710-4204-8055-ecea69f828ee" (UID: "45725226-1710-4204-8055-ecea69f828ee"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.635692 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" (UID: "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.635726 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-client-ca" (OuterVolumeSpecName: "client-ca") pod "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" (UID: "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.635824 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-config" (OuterVolumeSpecName: "config") pod "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" (UID: "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.636413 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-config" (OuterVolumeSpecName: "config") pod "45725226-1710-4204-8055-ecea69f828ee" (UID: "45725226-1710-4204-8055-ecea69f828ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.641613 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-kube-api-access-nx26v" (OuterVolumeSpecName: "kube-api-access-nx26v") pod "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" (UID: "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44"). InnerVolumeSpecName "kube-api-access-nx26v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.642246 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45725226-1710-4204-8055-ecea69f828ee-kube-api-access-4ssrd" (OuterVolumeSpecName: "kube-api-access-4ssrd") pod "45725226-1710-4204-8055-ecea69f828ee" (UID: "45725226-1710-4204-8055-ecea69f828ee"). InnerVolumeSpecName "kube-api-access-4ssrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.645938 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45725226-1710-4204-8055-ecea69f828ee-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "45725226-1710-4204-8055-ecea69f828ee" (UID: "45725226-1710-4204-8055-ecea69f828ee"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.651069 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" (UID: "7df9ce28-8006-4c4f-9d21-4c0ef83d2f44"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.736371 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.736408 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.736418 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/45725226-1710-4204-8055-ecea69f828ee-client-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.736427 4923 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-client-ca\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.736435 4923 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.736448 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.736459 4923 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45725226-1710-4204-8055-ecea69f828ee-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.737027 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nx26v\" (UniqueName: \"kubernetes.io/projected/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44-kube-api-access-nx26v\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.737068 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ssrd\" (UniqueName: \"kubernetes.io/projected/45725226-1710-4204-8055-ecea69f828ee-kube-api-access-4ssrd\") on node \"crc\" DevicePath \"\"" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.985656 4923 generic.go:334] "Generic (PLEG): container finished" podID="7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" containerID="0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f" exitCode=0 Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.985738 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.985736 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" event={"ID":"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44","Type":"ContainerDied","Data":"0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f"} Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.985908 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-tnbft" event={"ID":"7df9ce28-8006-4c4f-9d21-4c0ef83d2f44","Type":"ContainerDied","Data":"5311fe0dca51dc4392f58a2b7fa77e4b8a70f6e55e9de94dadbd9f9660ee0abb"} Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.985941 4923 scope.go:117] "RemoveContainer" containerID="0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f" Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.987860 4923 generic.go:334] "Generic (PLEG): container finished" podID="45725226-1710-4204-8055-ecea69f828ee" containerID="9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e" exitCode=0 Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.987890 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" event={"ID":"45725226-1710-4204-8055-ecea69f828ee","Type":"ContainerDied","Data":"9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e"} Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.987928 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" event={"ID":"45725226-1710-4204-8055-ecea69f828ee","Type":"ContainerDied","Data":"6b2f75cbad98fd90401800f66f8507673fc44ac6f1f00c94756370a0a2fb240b"} Oct 09 10:17:39 crc kubenswrapper[4923]: I1009 10:17:39.987974 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.010515 4923 scope.go:117] "RemoveContainer" containerID="0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f" Oct 09 10:17:40 crc kubenswrapper[4923]: E1009 10:17:40.013231 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f\": container with ID starting with 0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f not found: ID does not exist" containerID="0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.013296 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f"} err="failed to get container status \"0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f\": rpc error: code = NotFound desc = could not find container \"0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f\": container with ID starting with 0073317dd5fe3b504fd2691710d0411ce2d12aa1a474051db15b7b05961aca6f not found: ID does not exist" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.013367 4923 scope.go:117] "RemoveContainer" containerID="9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.026594 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tnbft"] Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.030416 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-tnbft"] Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.036199 4923 scope.go:117] "RemoveContainer" containerID="9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e" Oct 09 10:17:40 crc kubenswrapper[4923]: E1009 10:17:40.036888 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e\": container with ID starting with 9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e not found: ID does not exist" containerID="9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.036987 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e"} err="failed to get container status \"9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e\": rpc error: code = NotFound desc = could not find container \"9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e\": container with ID starting with 9e442198c06c38ea2b924881be6142d8de5968983fa9ab5494ae6cf6aa4d6b0e not found: ID does not exist" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.040461 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64"] Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.048117 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-d8h64"] Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.629308 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45725226-1710-4204-8055-ecea69f828ee" path="/var/lib/kubelet/pods/45725226-1710-4204-8055-ecea69f828ee/volumes" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.630041 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" path="/var/lib/kubelet/pods/7df9ce28-8006-4c4f-9d21-4c0ef83d2f44/volumes" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.821698 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-75c6887c96-26fnx"] Oct 09 10:17:40 crc kubenswrapper[4923]: E1009 10:17:40.822464 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" containerName="controller-manager" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.822485 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" containerName="controller-manager" Oct 09 10:17:40 crc kubenswrapper[4923]: E1009 10:17:40.822496 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45725226-1710-4204-8055-ecea69f828ee" containerName="route-controller-manager" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.822503 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="45725226-1710-4204-8055-ecea69f828ee" containerName="route-controller-manager" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.822610 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="45725226-1710-4204-8055-ecea69f828ee" containerName="route-controller-manager" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.822632 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df9ce28-8006-4c4f-9d21-4c0ef83d2f44" containerName="controller-manager" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.823118 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.839708 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.840146 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.840316 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.840462 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.842507 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.842932 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.848132 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.848359 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c6887c96-26fnx"] Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.875834 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8"] Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.877037 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.880400 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.880931 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.881074 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.881124 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.881238 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.881272 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.896669 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8"] Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.956226 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tdrh\" (UniqueName: \"kubernetes.io/projected/01bb65eb-52db-439d-ab2e-b62c3c57bafa-kube-api-access-2tdrh\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.956301 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01bb65eb-52db-439d-ab2e-b62c3c57bafa-config\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.956766 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01bb65eb-52db-439d-ab2e-b62c3c57bafa-serving-cert\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.956821 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/01bb65eb-52db-439d-ab2e-b62c3c57bafa-client-ca\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:40 crc kubenswrapper[4923]: I1009 10:17:40.956915 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/01bb65eb-52db-439d-ab2e-b62c3c57bafa-proxy-ca-bundles\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.058257 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01bb65eb-52db-439d-ab2e-b62c3c57bafa-config\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.058350 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e199d2ed-7976-49a2-bc96-b49f833ae2a0-serving-cert\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.058385 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg7ct\" (UniqueName: \"kubernetes.io/projected/e199d2ed-7976-49a2-bc96-b49f833ae2a0-kube-api-access-cg7ct\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.058464 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01bb65eb-52db-439d-ab2e-b62c3c57bafa-serving-cert\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.058491 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/01bb65eb-52db-439d-ab2e-b62c3c57bafa-client-ca\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.058527 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/01bb65eb-52db-439d-ab2e-b62c3c57bafa-proxy-ca-bundles\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.058709 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e199d2ed-7976-49a2-bc96-b49f833ae2a0-config\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.058785 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e199d2ed-7976-49a2-bc96-b49f833ae2a0-client-ca\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.058867 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tdrh\" (UniqueName: \"kubernetes.io/projected/01bb65eb-52db-439d-ab2e-b62c3c57bafa-kube-api-access-2tdrh\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.059813 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/01bb65eb-52db-439d-ab2e-b62c3c57bafa-proxy-ca-bundles\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.059833 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01bb65eb-52db-439d-ab2e-b62c3c57bafa-config\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.060281 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/01bb65eb-52db-439d-ab2e-b62c3c57bafa-client-ca\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.065643 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01bb65eb-52db-439d-ab2e-b62c3c57bafa-serving-cert\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.080004 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tdrh\" (UniqueName: \"kubernetes.io/projected/01bb65eb-52db-439d-ab2e-b62c3c57bafa-kube-api-access-2tdrh\") pod \"controller-manager-75c6887c96-26fnx\" (UID: \"01bb65eb-52db-439d-ab2e-b62c3c57bafa\") " pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.141209 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.159864 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e199d2ed-7976-49a2-bc96-b49f833ae2a0-config\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.159916 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e199d2ed-7976-49a2-bc96-b49f833ae2a0-client-ca\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.159957 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e199d2ed-7976-49a2-bc96-b49f833ae2a0-serving-cert\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.159979 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg7ct\" (UniqueName: \"kubernetes.io/projected/e199d2ed-7976-49a2-bc96-b49f833ae2a0-kube-api-access-cg7ct\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.161210 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e199d2ed-7976-49a2-bc96-b49f833ae2a0-client-ca\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.162095 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e199d2ed-7976-49a2-bc96-b49f833ae2a0-config\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.165018 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e199d2ed-7976-49a2-bc96-b49f833ae2a0-serving-cert\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.179318 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg7ct\" (UniqueName: \"kubernetes.io/projected/e199d2ed-7976-49a2-bc96-b49f833ae2a0-kube-api-access-cg7ct\") pod \"route-controller-manager-6589cd574-5k7v8\" (UID: \"e199d2ed-7976-49a2-bc96-b49f833ae2a0\") " pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.196501 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.449410 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-75c6887c96-26fnx"] Oct 09 10:17:41 crc kubenswrapper[4923]: I1009 10:17:41.513048 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8"] Oct 09 10:17:41 crc kubenswrapper[4923]: W1009 10:17:41.522066 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode199d2ed_7976_49a2_bc96_b49f833ae2a0.slice/crio-3472938a579a5ea2868fed138e5fc51e512f0406c7591a7a28833c12d10a5fe6 WatchSource:0}: Error finding container 3472938a579a5ea2868fed138e5fc51e512f0406c7591a7a28833c12d10a5fe6: Status 404 returned error can't find the container with id 3472938a579a5ea2868fed138e5fc51e512f0406c7591a7a28833c12d10a5fe6 Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.007267 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" event={"ID":"01bb65eb-52db-439d-ab2e-b62c3c57bafa","Type":"ContainerStarted","Data":"3691fc075856cee8eeabe498d77d79a2f220d5ea37703c2a1ee788ecb9105e02"} Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.007866 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.007885 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" event={"ID":"01bb65eb-52db-439d-ab2e-b62c3c57bafa","Type":"ContainerStarted","Data":"59c9e1738f7f7c2d88b1cdb432a52c1e473748c0cfe8ee752df5cc9a3bd56062"} Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.008693 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" event={"ID":"e199d2ed-7976-49a2-bc96-b49f833ae2a0","Type":"ContainerStarted","Data":"2b5017e15b2d95ce366919f10d24875e87db0cc5dd3bd7d4d4c8e7ab9f1fabc8"} Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.008740 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" event={"ID":"e199d2ed-7976-49a2-bc96-b49f833ae2a0","Type":"ContainerStarted","Data":"3472938a579a5ea2868fed138e5fc51e512f0406c7591a7a28833c12d10a5fe6"} Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.008781 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.021682 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.042621 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-75c6887c96-26fnx" podStartSLOduration=2.0425910800000002 podStartE2EDuration="2.04259108s" podCreationTimestamp="2025-10-09 10:17:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:17:42.040895553 +0000 UTC m=+748.109077319" watchObservedRunningTime="2025-10-09 10:17:42.04259108 +0000 UTC m=+748.110772846" Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.485463 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" Oct 09 10:17:42 crc kubenswrapper[4923]: I1009 10:17:42.522164 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6589cd574-5k7v8" podStartSLOduration=2.522130815 podStartE2EDuration="2.522130815s" podCreationTimestamp="2025-10-09 10:17:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:17:42.157494017 +0000 UTC m=+748.225675783" watchObservedRunningTime="2025-10-09 10:17:42.522130815 +0000 UTC m=+748.590312581" Oct 09 10:17:45 crc kubenswrapper[4923]: I1009 10:17:45.801372 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-b6c67695c-wqcnz" Oct 09 10:17:52 crc kubenswrapper[4923]: I1009 10:17:52.321935 4923 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 09 10:17:54 crc kubenswrapper[4923]: I1009 10:17:54.599864 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:17:54 crc kubenswrapper[4923]: I1009 10:17:54.600360 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:18:05 crc kubenswrapper[4923]: I1009 10:18:05.604265 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-59d99554d9-d78r5" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.320103 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-vsqk8"] Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.323276 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.326099 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.326147 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.332087 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-vnfh9" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.343916 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6"] Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.344977 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:06 crc kubenswrapper[4923]: W1009 10:18:06.347190 4923 reflector.go:561] object-"metallb-system"/"frr-k8s-webhook-server-cert": failed to list *v1.Secret: secrets "frr-k8s-webhook-server-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 09 10:18:06 crc kubenswrapper[4923]: E1009 10:18:06.347244 4923 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-webhook-server-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-webhook-server-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.362934 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6"] Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.432954 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-58t9k"] Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.434168 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.438083 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.438149 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.438201 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.438394 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-vbnn5" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.452944 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-fgmg8"] Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.454150 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.455718 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.471466 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-frr-conf\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.471523 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-reloader\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.471562 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/84d4e047-944f-41c7-859c-658a76ecc3be-metrics-certs\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.471584 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-frr-sockets\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.471594 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-fgmg8"] Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.471613 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/84d4e047-944f-41c7-859c-658a76ecc3be-frr-startup\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.471887 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzh4r\" (UniqueName: \"kubernetes.io/projected/b15e8e44-4023-48fe-8428-7b89368ada72-kube-api-access-mzh4r\") pod \"frr-k8s-webhook-server-64bf5d555-cp5k6\" (UID: \"b15e8e44-4023-48fe-8428-7b89368ada72\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.471920 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-metrics\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.472580 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvvfh\" (UniqueName: \"kubernetes.io/projected/84d4e047-944f-41c7-859c-658a76ecc3be-kube-api-access-rvvfh\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.472683 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b15e8e44-4023-48fe-8428-7b89368ada72-cert\") pod \"frr-k8s-webhook-server-64bf5d555-cp5k6\" (UID: \"b15e8e44-4023-48fe-8428-7b89368ada72\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574200 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-reloader\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574257 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctqkf\" (UniqueName: \"kubernetes.io/projected/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-kube-api-access-ctqkf\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574292 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70363f45-667f-4680-bf83-5358c618b73d-cert\") pod \"controller-68d546b9d8-fgmg8\" (UID: \"70363f45-667f-4680-bf83-5358c618b73d\") " pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574320 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70363f45-667f-4680-bf83-5358c618b73d-metrics-certs\") pod \"controller-68d546b9d8-fgmg8\" (UID: \"70363f45-667f-4680-bf83-5358c618b73d\") " pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574349 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-metallb-excludel2\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574372 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-metrics-certs\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574394 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/84d4e047-944f-41c7-859c-658a76ecc3be-metrics-certs\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574564 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-frr-sockets\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574607 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/84d4e047-944f-41c7-859c-658a76ecc3be-frr-startup\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574646 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzh4r\" (UniqueName: \"kubernetes.io/projected/b15e8e44-4023-48fe-8428-7b89368ada72-kube-api-access-mzh4r\") pod \"frr-k8s-webhook-server-64bf5d555-cp5k6\" (UID: \"b15e8e44-4023-48fe-8428-7b89368ada72\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574673 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-metrics\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574700 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvvfh\" (UniqueName: \"kubernetes.io/projected/84d4e047-944f-41c7-859c-658a76ecc3be-kube-api-access-rvvfh\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574725 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b15e8e44-4023-48fe-8428-7b89368ada72-cert\") pod \"frr-k8s-webhook-server-64bf5d555-cp5k6\" (UID: \"b15e8e44-4023-48fe-8428-7b89368ada72\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574771 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-memberlist\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574819 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfm6r\" (UniqueName: \"kubernetes.io/projected/70363f45-667f-4680-bf83-5358c618b73d-kube-api-access-sfm6r\") pod \"controller-68d546b9d8-fgmg8\" (UID: \"70363f45-667f-4680-bf83-5358c618b73d\") " pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574845 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-frr-conf\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.574990 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-reloader\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.575197 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-frr-sockets\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.575507 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-frr-conf\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.575786 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/84d4e047-944f-41c7-859c-658a76ecc3be-metrics\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.576033 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/84d4e047-944f-41c7-859c-658a76ecc3be-frr-startup\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.582775 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/84d4e047-944f-41c7-859c-658a76ecc3be-metrics-certs\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.594977 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvvfh\" (UniqueName: \"kubernetes.io/projected/84d4e047-944f-41c7-859c-658a76ecc3be-kube-api-access-rvvfh\") pod \"frr-k8s-vsqk8\" (UID: \"84d4e047-944f-41c7-859c-658a76ecc3be\") " pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.606373 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzh4r\" (UniqueName: \"kubernetes.io/projected/b15e8e44-4023-48fe-8428-7b89368ada72-kube-api-access-mzh4r\") pod \"frr-k8s-webhook-server-64bf5d555-cp5k6\" (UID: \"b15e8e44-4023-48fe-8428-7b89368ada72\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.642349 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.677455 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-memberlist\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.677540 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sfm6r\" (UniqueName: \"kubernetes.io/projected/70363f45-667f-4680-bf83-5358c618b73d-kube-api-access-sfm6r\") pod \"controller-68d546b9d8-fgmg8\" (UID: \"70363f45-667f-4680-bf83-5358c618b73d\") " pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.677583 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctqkf\" (UniqueName: \"kubernetes.io/projected/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-kube-api-access-ctqkf\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.677607 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70363f45-667f-4680-bf83-5358c618b73d-cert\") pod \"controller-68d546b9d8-fgmg8\" (UID: \"70363f45-667f-4680-bf83-5358c618b73d\") " pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.677633 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70363f45-667f-4680-bf83-5358c618b73d-metrics-certs\") pod \"controller-68d546b9d8-fgmg8\" (UID: \"70363f45-667f-4680-bf83-5358c618b73d\") " pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.677666 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-metallb-excludel2\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.677695 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-metrics-certs\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: E1009 10:18:06.678731 4923 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 09 10:18:06 crc kubenswrapper[4923]: E1009 10:18:06.678918 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-memberlist podName:e3ed7aa1-9fa9-4dd6-9701-3845073d3422 nodeName:}" failed. No retries permitted until 2025-10-09 10:18:07.178824625 +0000 UTC m=+773.247006381 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-memberlist") pod "speaker-58t9k" (UID: "e3ed7aa1-9fa9-4dd6-9701-3845073d3422") : secret "metallb-memberlist" not found Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.690415 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-metallb-excludel2\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.690926 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-metrics-certs\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.693148 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/70363f45-667f-4680-bf83-5358c618b73d-cert\") pod \"controller-68d546b9d8-fgmg8\" (UID: \"70363f45-667f-4680-bf83-5358c618b73d\") " pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.694665 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70363f45-667f-4680-bf83-5358c618b73d-metrics-certs\") pod \"controller-68d546b9d8-fgmg8\" (UID: \"70363f45-667f-4680-bf83-5358c618b73d\") " pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.710209 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sfm6r\" (UniqueName: \"kubernetes.io/projected/70363f45-667f-4680-bf83-5358c618b73d-kube-api-access-sfm6r\") pod \"controller-68d546b9d8-fgmg8\" (UID: \"70363f45-667f-4680-bf83-5358c618b73d\") " pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.718737 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctqkf\" (UniqueName: \"kubernetes.io/projected/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-kube-api-access-ctqkf\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:06 crc kubenswrapper[4923]: I1009 10:18:06.769966 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:07 crc kubenswrapper[4923]: I1009 10:18:07.185378 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-memberlist\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:07 crc kubenswrapper[4923]: E1009 10:18:07.185579 4923 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 09 10:18:07 crc kubenswrapper[4923]: E1009 10:18:07.185987 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-memberlist podName:e3ed7aa1-9fa9-4dd6-9701-3845073d3422 nodeName:}" failed. No retries permitted until 2025-10-09 10:18:08.185963323 +0000 UTC m=+774.254145079 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-memberlist") pod "speaker-58t9k" (UID: "e3ed7aa1-9fa9-4dd6-9701-3845073d3422") : secret "metallb-memberlist" not found Oct 09 10:18:07 crc kubenswrapper[4923]: I1009 10:18:07.192245 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-fgmg8"] Oct 09 10:18:07 crc kubenswrapper[4923]: W1009 10:18:07.201504 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70363f45_667f_4680_bf83_5358c618b73d.slice/crio-71e5aa7455e2b3e27f01c2da882e894085cc068a900e7fbbff992a1f28a29dd7 WatchSource:0}: Error finding container 71e5aa7455e2b3e27f01c2da882e894085cc068a900e7fbbff992a1f28a29dd7: Status 404 returned error can't find the container with id 71e5aa7455e2b3e27f01c2da882e894085cc068a900e7fbbff992a1f28a29dd7 Oct 09 10:18:07 crc kubenswrapper[4923]: I1009 10:18:07.205248 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerStarted","Data":"55588dbbb5a974da2edcb21cae480560b97eb0271fe5cd5a26af127d034e59c3"} Oct 09 10:18:07 crc kubenswrapper[4923]: E1009 10:18:07.576340 4923 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: failed to sync secret cache: timed out waiting for the condition Oct 09 10:18:07 crc kubenswrapper[4923]: E1009 10:18:07.576898 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b15e8e44-4023-48fe-8428-7b89368ada72-cert podName:b15e8e44-4023-48fe-8428-7b89368ada72 nodeName:}" failed. No retries permitted until 2025-10-09 10:18:08.076873607 +0000 UTC m=+774.145055363 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b15e8e44-4023-48fe-8428-7b89368ada72-cert") pod "frr-k8s-webhook-server-64bf5d555-cp5k6" (UID: "b15e8e44-4023-48fe-8428-7b89368ada72") : failed to sync secret cache: timed out waiting for the condition Oct 09 10:18:07 crc kubenswrapper[4923]: I1009 10:18:07.947606 4923 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.098828 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b15e8e44-4023-48fe-8428-7b89368ada72-cert\") pod \"frr-k8s-webhook-server-64bf5d555-cp5k6\" (UID: \"b15e8e44-4023-48fe-8428-7b89368ada72\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.111511 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b15e8e44-4023-48fe-8428-7b89368ada72-cert\") pod \"frr-k8s-webhook-server-64bf5d555-cp5k6\" (UID: \"b15e8e44-4023-48fe-8428-7b89368ada72\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.158934 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.199630 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-memberlist\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.204875 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/e3ed7aa1-9fa9-4dd6-9701-3845073d3422-memberlist\") pod \"speaker-58t9k\" (UID: \"e3ed7aa1-9fa9-4dd6-9701-3845073d3422\") " pod="metallb-system/speaker-58t9k" Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.212123 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-fgmg8" event={"ID":"70363f45-667f-4680-bf83-5358c618b73d","Type":"ContainerStarted","Data":"5d27fade6ca801b90648f89e29dd2de56021798e297062aa0dea2dc72aa185c0"} Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.212173 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-fgmg8" event={"ID":"70363f45-667f-4680-bf83-5358c618b73d","Type":"ContainerStarted","Data":"76a202b8485754295af275fc0f51535b5321c4a71a0037e682e6cab90e8b892b"} Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.212186 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-fgmg8" event={"ID":"70363f45-667f-4680-bf83-5358c618b73d","Type":"ContainerStarted","Data":"71e5aa7455e2b3e27f01c2da882e894085cc068a900e7fbbff992a1f28a29dd7"} Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.212283 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.231965 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-fgmg8" podStartSLOduration=2.231939815 podStartE2EDuration="2.231939815s" podCreationTimestamp="2025-10-09 10:18:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:18:08.226633596 +0000 UTC m=+774.294815352" watchObservedRunningTime="2025-10-09 10:18:08.231939815 +0000 UTC m=+774.300121571" Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.248112 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-58t9k" Oct 09 10:18:08 crc kubenswrapper[4923]: W1009 10:18:08.274529 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode3ed7aa1_9fa9_4dd6_9701_3845073d3422.slice/crio-c6ea1a7eeebca0fa7c1a612d3d514551b17c44eabf93d9a0a70af522b353d305 WatchSource:0}: Error finding container c6ea1a7eeebca0fa7c1a612d3d514551b17c44eabf93d9a0a70af522b353d305: Status 404 returned error can't find the container with id c6ea1a7eeebca0fa7c1a612d3d514551b17c44eabf93d9a0a70af522b353d305 Oct 09 10:18:08 crc kubenswrapper[4923]: I1009 10:18:08.627051 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6"] Oct 09 10:18:08 crc kubenswrapper[4923]: W1009 10:18:08.635920 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb15e8e44_4023_48fe_8428_7b89368ada72.slice/crio-19314b5fa425a5fce4f219f1c43a9e7f59a0ecb6f9c177cef54240eba9fa849a WatchSource:0}: Error finding container 19314b5fa425a5fce4f219f1c43a9e7f59a0ecb6f9c177cef54240eba9fa849a: Status 404 returned error can't find the container with id 19314b5fa425a5fce4f219f1c43a9e7f59a0ecb6f9c177cef54240eba9fa849a Oct 09 10:18:09 crc kubenswrapper[4923]: I1009 10:18:09.222735 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" event={"ID":"b15e8e44-4023-48fe-8428-7b89368ada72","Type":"ContainerStarted","Data":"19314b5fa425a5fce4f219f1c43a9e7f59a0ecb6f9c177cef54240eba9fa849a"} Oct 09 10:18:09 crc kubenswrapper[4923]: I1009 10:18:09.232245 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-58t9k" event={"ID":"e3ed7aa1-9fa9-4dd6-9701-3845073d3422","Type":"ContainerStarted","Data":"0c6d93d12b9dd36fdcd02ec7ec58b51967f7da7bb3f54a2bb00a4a7154642a64"} Oct 09 10:18:09 crc kubenswrapper[4923]: I1009 10:18:09.232297 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-58t9k" event={"ID":"e3ed7aa1-9fa9-4dd6-9701-3845073d3422","Type":"ContainerStarted","Data":"b03ededbfb779c2b58214dd06889c9bf29aef922a1a05bfc7035924e8d7ba571"} Oct 09 10:18:09 crc kubenswrapper[4923]: I1009 10:18:09.232307 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-58t9k" event={"ID":"e3ed7aa1-9fa9-4dd6-9701-3845073d3422","Type":"ContainerStarted","Data":"c6ea1a7eeebca0fa7c1a612d3d514551b17c44eabf93d9a0a70af522b353d305"} Oct 09 10:18:09 crc kubenswrapper[4923]: I1009 10:18:09.232933 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-58t9k" Oct 09 10:18:09 crc kubenswrapper[4923]: I1009 10:18:09.253491 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-58t9k" podStartSLOduration=3.253463182 podStartE2EDuration="3.253463182s" podCreationTimestamp="2025-10-09 10:18:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:18:09.2526561 +0000 UTC m=+775.320837856" watchObservedRunningTime="2025-10-09 10:18:09.253463182 +0000 UTC m=+775.321644938" Oct 09 10:18:15 crc kubenswrapper[4923]: I1009 10:18:15.287625 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" event={"ID":"b15e8e44-4023-48fe-8428-7b89368ada72","Type":"ContainerStarted","Data":"9e427f7f94f9427622a3301077ae7117f7ff3a3370d3f0459e03c54393af641c"} Oct 09 10:18:15 crc kubenswrapper[4923]: I1009 10:18:15.288721 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:15 crc kubenswrapper[4923]: I1009 10:18:15.289720 4923 generic.go:334] "Generic (PLEG): container finished" podID="84d4e047-944f-41c7-859c-658a76ecc3be" containerID="cd537bbff98bac548bc9e8477828c390dac10710fbab1ebd3c8b5b9ebbb7ce9c" exitCode=0 Oct 09 10:18:15 crc kubenswrapper[4923]: I1009 10:18:15.289795 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerDied","Data":"cd537bbff98bac548bc9e8477828c390dac10710fbab1ebd3c8b5b9ebbb7ce9c"} Oct 09 10:18:15 crc kubenswrapper[4923]: I1009 10:18:15.313332 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" podStartSLOduration=3.614639914 podStartE2EDuration="9.313308217s" podCreationTimestamp="2025-10-09 10:18:06 +0000 UTC" firstStartedPulling="2025-10-09 10:18:08.640722229 +0000 UTC m=+774.708903985" lastFinishedPulling="2025-10-09 10:18:14.339390492 +0000 UTC m=+780.407572288" observedRunningTime="2025-10-09 10:18:15.309081358 +0000 UTC m=+781.377263144" watchObservedRunningTime="2025-10-09 10:18:15.313308217 +0000 UTC m=+781.381489983" Oct 09 10:18:16 crc kubenswrapper[4923]: I1009 10:18:16.303390 4923 generic.go:334] "Generic (PLEG): container finished" podID="84d4e047-944f-41c7-859c-658a76ecc3be" containerID="9415267dc5830697ad02297e63d14bf7f6a23b1c35b3506ef813da301b964e98" exitCode=0 Oct 09 10:18:16 crc kubenswrapper[4923]: I1009 10:18:16.303471 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerDied","Data":"9415267dc5830697ad02297e63d14bf7f6a23b1c35b3506ef813da301b964e98"} Oct 09 10:18:17 crc kubenswrapper[4923]: I1009 10:18:17.314981 4923 generic.go:334] "Generic (PLEG): container finished" podID="84d4e047-944f-41c7-859c-658a76ecc3be" containerID="081b0fcdbbe0ad493ff162d509cfd9a03d3543f7c1cd6b922f566b59d28282e3" exitCode=0 Oct 09 10:18:17 crc kubenswrapper[4923]: I1009 10:18:17.315097 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerDied","Data":"081b0fcdbbe0ad493ff162d509cfd9a03d3543f7c1cd6b922f566b59d28282e3"} Oct 09 10:18:18 crc kubenswrapper[4923]: I1009 10:18:18.253694 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-58t9k" Oct 09 10:18:18 crc kubenswrapper[4923]: I1009 10:18:18.329229 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerStarted","Data":"4751630123e58eb09f97914edab0b9491b3b39aeb374a2f6308e9f37d8855e5f"} Oct 09 10:18:18 crc kubenswrapper[4923]: I1009 10:18:18.329284 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerStarted","Data":"5b006f71d3fd1fa7ed83b8e8bde1453606f451147bf4a30937d48f0d1cbd024c"} Oct 09 10:18:18 crc kubenswrapper[4923]: I1009 10:18:18.329297 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerStarted","Data":"247c9d2777c0da7c3ac64a1363d902666b3ad77f011a92d6f1cbea6acf9f73d7"} Oct 09 10:18:18 crc kubenswrapper[4923]: I1009 10:18:18.329307 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerStarted","Data":"16e1f531d8249144f83498b9d7c3b47d91cc11e4369a35ccd95159cf7ef280ed"} Oct 09 10:18:18 crc kubenswrapper[4923]: I1009 10:18:18.329320 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerStarted","Data":"42cdcd6d937098aded88eb46d4cda5ee80d9a92bddb82858ec870a729c39037b"} Oct 09 10:18:19 crc kubenswrapper[4923]: I1009 10:18:19.341490 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-vsqk8" event={"ID":"84d4e047-944f-41c7-859c-658a76ecc3be","Type":"ContainerStarted","Data":"1e5e20fac018e781099fbd28b02e10cbee35d56c7910c419f7b4f7fc4bf11298"} Oct 09 10:18:19 crc kubenswrapper[4923]: I1009 10:18:19.341818 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:19 crc kubenswrapper[4923]: I1009 10:18:19.371253 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-vsqk8" podStartSLOduration=5.861519284 podStartE2EDuration="13.371224087s" podCreationTimestamp="2025-10-09 10:18:06 +0000 UTC" firstStartedPulling="2025-10-09 10:18:06.828953259 +0000 UTC m=+772.897135015" lastFinishedPulling="2025-10-09 10:18:14.338658022 +0000 UTC m=+780.406839818" observedRunningTime="2025-10-09 10:18:19.368037888 +0000 UTC m=+785.436219644" watchObservedRunningTime="2025-10-09 10:18:19.371224087 +0000 UTC m=+785.439405843" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.021922 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-fgscm"] Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.023238 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fgscm" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.030010 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fgscm"] Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.032464 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.032658 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.032899 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-8vhs9" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.104416 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swvv5\" (UniqueName: \"kubernetes.io/projected/11141990-269a-4525-b145-a3f9efdc39bb-kube-api-access-swvv5\") pod \"openstack-operator-index-fgscm\" (UID: \"11141990-269a-4525-b145-a3f9efdc39bb\") " pod="openstack-operators/openstack-operator-index-fgscm" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.206179 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swvv5\" (UniqueName: \"kubernetes.io/projected/11141990-269a-4525-b145-a3f9efdc39bb-kube-api-access-swvv5\") pod \"openstack-operator-index-fgscm\" (UID: \"11141990-269a-4525-b145-a3f9efdc39bb\") " pod="openstack-operators/openstack-operator-index-fgscm" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.227559 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swvv5\" (UniqueName: \"kubernetes.io/projected/11141990-269a-4525-b145-a3f9efdc39bb-kube-api-access-swvv5\") pod \"openstack-operator-index-fgscm\" (UID: \"11141990-269a-4525-b145-a3f9efdc39bb\") " pod="openstack-operators/openstack-operator-index-fgscm" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.346498 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fgscm" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.643566 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.700404 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:21 crc kubenswrapper[4923]: I1009 10:18:21.804480 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-fgscm"] Oct 09 10:18:22 crc kubenswrapper[4923]: I1009 10:18:22.375989 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fgscm" event={"ID":"11141990-269a-4525-b145-a3f9efdc39bb","Type":"ContainerStarted","Data":"6cac9da63558c1731cfe03f0855dbaa0764c7eb04b934daec31bf7692545a01a"} Oct 09 10:18:23 crc kubenswrapper[4923]: I1009 10:18:23.384769 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fgscm" event={"ID":"11141990-269a-4525-b145-a3f9efdc39bb","Type":"ContainerStarted","Data":"a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce"} Oct 09 10:18:23 crc kubenswrapper[4923]: I1009 10:18:23.408465 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-fgscm" podStartSLOduration=1.53050736 podStartE2EDuration="2.408436027s" podCreationTimestamp="2025-10-09 10:18:21 +0000 UTC" firstStartedPulling="2025-10-09 10:18:21.807235923 +0000 UTC m=+787.875417719" lastFinishedPulling="2025-10-09 10:18:22.68516461 +0000 UTC m=+788.753346386" observedRunningTime="2025-10-09 10:18:23.404150298 +0000 UTC m=+789.472332094" watchObservedRunningTime="2025-10-09 10:18:23.408436027 +0000 UTC m=+789.476617793" Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.201736 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fgscm"] Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.599777 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.601006 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.613097 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.613853 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af930c04f923630accaccaa4b93be2a9110fb2ade8cd00a8a79620f54fdb4120"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.613972 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://af930c04f923630accaccaa4b93be2a9110fb2ade8cd00a8a79620f54fdb4120" gracePeriod=600 Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.805435 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-bbtq2"] Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.806238 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bbtq2" Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.817833 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bbtq2"] Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.870465 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfkfv\" (UniqueName: \"kubernetes.io/projected/8ba088a5-a6d3-48ab-8966-efd7be3f7981-kube-api-access-cfkfv\") pod \"openstack-operator-index-bbtq2\" (UID: \"8ba088a5-a6d3-48ab-8966-efd7be3f7981\") " pod="openstack-operators/openstack-operator-index-bbtq2" Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.971663 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfkfv\" (UniqueName: \"kubernetes.io/projected/8ba088a5-a6d3-48ab-8966-efd7be3f7981-kube-api-access-cfkfv\") pod \"openstack-operator-index-bbtq2\" (UID: \"8ba088a5-a6d3-48ab-8966-efd7be3f7981\") " pod="openstack-operators/openstack-operator-index-bbtq2" Oct 09 10:18:24 crc kubenswrapper[4923]: I1009 10:18:24.996900 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfkfv\" (UniqueName: \"kubernetes.io/projected/8ba088a5-a6d3-48ab-8966-efd7be3f7981-kube-api-access-cfkfv\") pod \"openstack-operator-index-bbtq2\" (UID: \"8ba088a5-a6d3-48ab-8966-efd7be3f7981\") " pod="openstack-operators/openstack-operator-index-bbtq2" Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.146086 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-bbtq2" Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.414918 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="af930c04f923630accaccaa4b93be2a9110fb2ade8cd00a8a79620f54fdb4120" exitCode=0 Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.415217 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-fgscm" podUID="11141990-269a-4525-b145-a3f9efdc39bb" containerName="registry-server" containerID="cri-o://a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce" gracePeriod=2 Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.416142 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"af930c04f923630accaccaa4b93be2a9110fb2ade8cd00a8a79620f54fdb4120"} Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.416185 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"0371acf1b4e04ebc9cc4694cbecac34256a8ea0829820a4e57e504509cef2cbf"} Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.416211 4923 scope.go:117] "RemoveContainer" containerID="15e3211b265dc1c111dcb9faf998805ee133fefab85afa177cdecf472b4fa20c" Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.692696 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-bbtq2"] Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.812937 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fgscm" Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.885025 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swvv5\" (UniqueName: \"kubernetes.io/projected/11141990-269a-4525-b145-a3f9efdc39bb-kube-api-access-swvv5\") pod \"11141990-269a-4525-b145-a3f9efdc39bb\" (UID: \"11141990-269a-4525-b145-a3f9efdc39bb\") " Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.893036 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11141990-269a-4525-b145-a3f9efdc39bb-kube-api-access-swvv5" (OuterVolumeSpecName: "kube-api-access-swvv5") pod "11141990-269a-4525-b145-a3f9efdc39bb" (UID: "11141990-269a-4525-b145-a3f9efdc39bb"). InnerVolumeSpecName "kube-api-access-swvv5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:18:25 crc kubenswrapper[4923]: I1009 10:18:25.988344 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swvv5\" (UniqueName: \"kubernetes.io/projected/11141990-269a-4525-b145-a3f9efdc39bb-kube-api-access-swvv5\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.425554 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bbtq2" event={"ID":"8ba088a5-a6d3-48ab-8966-efd7be3f7981","Type":"ContainerStarted","Data":"5b26b1de71f589c2ef698b045479f49c4195ac0d6ff06545bdc0617b16412046"} Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.425934 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-bbtq2" event={"ID":"8ba088a5-a6d3-48ab-8966-efd7be3f7981","Type":"ContainerStarted","Data":"16e192c8c21812df25b92582ac97d5411dfd89c788c90859f7d148f287e63450"} Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.429106 4923 generic.go:334] "Generic (PLEG): container finished" podID="11141990-269a-4525-b145-a3f9efdc39bb" containerID="a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce" exitCode=0 Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.429138 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fgscm" event={"ID":"11141990-269a-4525-b145-a3f9efdc39bb","Type":"ContainerDied","Data":"a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce"} Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.429157 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-fgscm" event={"ID":"11141990-269a-4525-b145-a3f9efdc39bb","Type":"ContainerDied","Data":"6cac9da63558c1731cfe03f0855dbaa0764c7eb04b934daec31bf7692545a01a"} Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.429175 4923 scope.go:117] "RemoveContainer" containerID="a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce" Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.429261 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-fgscm" Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.448623 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-bbtq2" podStartSLOduration=2.0516790560000002 podStartE2EDuration="2.448598057s" podCreationTimestamp="2025-10-09 10:18:24 +0000 UTC" firstStartedPulling="2025-10-09 10:18:25.709506656 +0000 UTC m=+791.777688412" lastFinishedPulling="2025-10-09 10:18:26.106425657 +0000 UTC m=+792.174607413" observedRunningTime="2025-10-09 10:18:26.442463054 +0000 UTC m=+792.510644810" watchObservedRunningTime="2025-10-09 10:18:26.448598057 +0000 UTC m=+792.516779813" Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.461036 4923 scope.go:117] "RemoveContainer" containerID="a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce" Oct 09 10:18:26 crc kubenswrapper[4923]: E1009 10:18:26.461856 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce\": container with ID starting with a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce not found: ID does not exist" containerID="a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce" Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.461900 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce"} err="failed to get container status \"a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce\": rpc error: code = NotFound desc = could not find container \"a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce\": container with ID starting with a746d261fd9820c237561c5a233f114c904b599776a7b8eb1c5e3f95eae257ce not found: ID does not exist" Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.462533 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-fgscm"] Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.466155 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-fgscm"] Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.614481 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11141990-269a-4525-b145-a3f9efdc39bb" path="/var/lib/kubelet/pods/11141990-269a-4525-b145-a3f9efdc39bb/volumes" Oct 09 10:18:26 crc kubenswrapper[4923]: I1009 10:18:26.776215 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-fgmg8" Oct 09 10:18:28 crc kubenswrapper[4923]: I1009 10:18:28.167614 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-cp5k6" Oct 09 10:18:35 crc kubenswrapper[4923]: I1009 10:18:35.147837 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-bbtq2" Oct 09 10:18:35 crc kubenswrapper[4923]: I1009 10:18:35.148683 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-bbtq2" Oct 09 10:18:35 crc kubenswrapper[4923]: I1009 10:18:35.190520 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-bbtq2" Oct 09 10:18:35 crc kubenswrapper[4923]: I1009 10:18:35.531822 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-bbtq2" Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.652591 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-vsqk8" Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.856080 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp"] Oct 09 10:18:36 crc kubenswrapper[4923]: E1009 10:18:36.856479 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11141990-269a-4525-b145-a3f9efdc39bb" containerName="registry-server" Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.856506 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="11141990-269a-4525-b145-a3f9efdc39bb" containerName="registry-server" Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.856683 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="11141990-269a-4525-b145-a3f9efdc39bb" containerName="registry-server" Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.857895 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.865727 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-452cl" Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.871387 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp"] Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.952055 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-flw9d\" (UniqueName: \"kubernetes.io/projected/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-kube-api-access-flw9d\") pod \"184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.952100 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-util\") pod \"184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:36 crc kubenswrapper[4923]: I1009 10:18:36.952179 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-bundle\") pod \"184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:37 crc kubenswrapper[4923]: I1009 10:18:37.053219 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-bundle\") pod \"184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:37 crc kubenswrapper[4923]: I1009 10:18:37.053333 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-flw9d\" (UniqueName: \"kubernetes.io/projected/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-kube-api-access-flw9d\") pod \"184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:37 crc kubenswrapper[4923]: I1009 10:18:37.053377 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-util\") pod \"184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:37 crc kubenswrapper[4923]: I1009 10:18:37.053966 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-bundle\") pod \"184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:37 crc kubenswrapper[4923]: I1009 10:18:37.054200 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-util\") pod \"184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:37 crc kubenswrapper[4923]: I1009 10:18:37.077073 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-flw9d\" (UniqueName: \"kubernetes.io/projected/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-kube-api-access-flw9d\") pod \"184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:37 crc kubenswrapper[4923]: I1009 10:18:37.225742 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:37 crc kubenswrapper[4923]: I1009 10:18:37.672761 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp"] Oct 09 10:18:37 crc kubenswrapper[4923]: W1009 10:18:37.686440 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1a2f9a94_3b55_48ee_aa73_d22e8bde118c.slice/crio-ecc991c4a8a3b5d4bc45d25f1dcc0a66f1889af1784105744a832791ee02dfdf WatchSource:0}: Error finding container ecc991c4a8a3b5d4bc45d25f1dcc0a66f1889af1784105744a832791ee02dfdf: Status 404 returned error can't find the container with id ecc991c4a8a3b5d4bc45d25f1dcc0a66f1889af1784105744a832791ee02dfdf Oct 09 10:18:38 crc kubenswrapper[4923]: I1009 10:18:38.535434 4923 generic.go:334] "Generic (PLEG): container finished" podID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerID="fed0cfd2b39013388129052cd03c1cf3ca06af90bb8f0454e16e92b8d24f3c6d" exitCode=0 Oct 09 10:18:38 crc kubenswrapper[4923]: I1009 10:18:38.535777 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" event={"ID":"1a2f9a94-3b55-48ee-aa73-d22e8bde118c","Type":"ContainerDied","Data":"fed0cfd2b39013388129052cd03c1cf3ca06af90bb8f0454e16e92b8d24f3c6d"} Oct 09 10:18:38 crc kubenswrapper[4923]: I1009 10:18:38.535806 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" event={"ID":"1a2f9a94-3b55-48ee-aa73-d22e8bde118c","Type":"ContainerStarted","Data":"ecc991c4a8a3b5d4bc45d25f1dcc0a66f1889af1784105744a832791ee02dfdf"} Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.236975 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q5954"] Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.239632 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5954"] Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.239829 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.387691 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-catalog-content\") pod \"redhat-marketplace-q5954\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.387827 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-utilities\") pod \"redhat-marketplace-q5954\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.387934 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpnbz\" (UniqueName: \"kubernetes.io/projected/17c9e5de-ac30-49e6-b6e4-46507b57b70d-kube-api-access-tpnbz\") pod \"redhat-marketplace-q5954\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.488842 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-catalog-content\") pod \"redhat-marketplace-q5954\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.488907 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-utilities\") pod \"redhat-marketplace-q5954\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.488976 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpnbz\" (UniqueName: \"kubernetes.io/projected/17c9e5de-ac30-49e6-b6e4-46507b57b70d-kube-api-access-tpnbz\") pod \"redhat-marketplace-q5954\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.489732 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-catalog-content\") pod \"redhat-marketplace-q5954\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.490038 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-utilities\") pod \"redhat-marketplace-q5954\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.518905 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpnbz\" (UniqueName: \"kubernetes.io/projected/17c9e5de-ac30-49e6-b6e4-46507b57b70d-kube-api-access-tpnbz\") pod \"redhat-marketplace-q5954\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.546894 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" event={"ID":"1a2f9a94-3b55-48ee-aa73-d22e8bde118c","Type":"ContainerStarted","Data":"63fac0fc6c79c4a778bb7a99579f0dbc03462b46acb73e47fcbdc51830e95552"} Oct 09 10:18:39 crc kubenswrapper[4923]: I1009 10:18:39.580358 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.038990 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5954"] Oct 09 10:18:40 crc kubenswrapper[4923]: W1009 10:18:40.048195 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17c9e5de_ac30_49e6_b6e4_46507b57b70d.slice/crio-b4488e4491ea4a678d3812d7c4389f72f382c2704863adbe7aafe51ade35cbef WatchSource:0}: Error finding container b4488e4491ea4a678d3812d7c4389f72f382c2704863adbe7aafe51ade35cbef: Status 404 returned error can't find the container with id b4488e4491ea4a678d3812d7c4389f72f382c2704863adbe7aafe51ade35cbef Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.203489 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wp4hl"] Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.205190 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.215143 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wp4hl"] Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.399517 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drk4k\" (UniqueName: \"kubernetes.io/projected/ecb6df4f-286a-4725-af44-c114db0a5e80-kube-api-access-drk4k\") pod \"redhat-operators-wp4hl\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.399666 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-utilities\") pod \"redhat-operators-wp4hl\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.399703 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-catalog-content\") pod \"redhat-operators-wp4hl\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.500949 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-catalog-content\") pod \"redhat-operators-wp4hl\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.501062 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drk4k\" (UniqueName: \"kubernetes.io/projected/ecb6df4f-286a-4725-af44-c114db0a5e80-kube-api-access-drk4k\") pod \"redhat-operators-wp4hl\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.501160 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-utilities\") pod \"redhat-operators-wp4hl\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.501616 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-catalog-content\") pod \"redhat-operators-wp4hl\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.501786 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-utilities\") pod \"redhat-operators-wp4hl\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.533887 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drk4k\" (UniqueName: \"kubernetes.io/projected/ecb6df4f-286a-4725-af44-c114db0a5e80-kube-api-access-drk4k\") pod \"redhat-operators-wp4hl\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.553953 4923 generic.go:334] "Generic (PLEG): container finished" podID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerID="e1581c9d2d2d252d94d02767378f06ba73ac9f135d07fdf220b1d53c7e081a6c" exitCode=0 Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.554039 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5954" event={"ID":"17c9e5de-ac30-49e6-b6e4-46507b57b70d","Type":"ContainerDied","Data":"e1581c9d2d2d252d94d02767378f06ba73ac9f135d07fdf220b1d53c7e081a6c"} Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.554076 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5954" event={"ID":"17c9e5de-ac30-49e6-b6e4-46507b57b70d","Type":"ContainerStarted","Data":"b4488e4491ea4a678d3812d7c4389f72f382c2704863adbe7aafe51ade35cbef"} Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.557872 4923 generic.go:334] "Generic (PLEG): container finished" podID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerID="63fac0fc6c79c4a778bb7a99579f0dbc03462b46acb73e47fcbdc51830e95552" exitCode=0 Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.557949 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" event={"ID":"1a2f9a94-3b55-48ee-aa73-d22e8bde118c","Type":"ContainerDied","Data":"63fac0fc6c79c4a778bb7a99579f0dbc03462b46acb73e47fcbdc51830e95552"} Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.562438 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:40 crc kubenswrapper[4923]: I1009 10:18:40.800714 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wp4hl"] Oct 09 10:18:40 crc kubenswrapper[4923]: W1009 10:18:40.809885 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podecb6df4f_286a_4725_af44_c114db0a5e80.slice/crio-2aaaa23cd615ce4383c5754f239e557b90e98be8abe92927aa3cb9ab15d59f8f WatchSource:0}: Error finding container 2aaaa23cd615ce4383c5754f239e557b90e98be8abe92927aa3cb9ab15d59f8f: Status 404 returned error can't find the container with id 2aaaa23cd615ce4383c5754f239e557b90e98be8abe92927aa3cb9ab15d59f8f Oct 09 10:18:41 crc kubenswrapper[4923]: I1009 10:18:41.570745 4923 generic.go:334] "Generic (PLEG): container finished" podID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerID="7d7a2e175884068dfadcecb77c3636618ae345e353e4048ef42b0bb95a16f669" exitCode=0 Oct 09 10:18:41 crc kubenswrapper[4923]: I1009 10:18:41.570925 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" event={"ID":"1a2f9a94-3b55-48ee-aa73-d22e8bde118c","Type":"ContainerDied","Data":"7d7a2e175884068dfadcecb77c3636618ae345e353e4048ef42b0bb95a16f669"} Oct 09 10:18:41 crc kubenswrapper[4923]: I1009 10:18:41.573091 4923 generic.go:334] "Generic (PLEG): container finished" podID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerID="14de722b2c88189f9520ff9e4472600ad926f97b972a95f0c750503f635ad9dc" exitCode=0 Oct 09 10:18:41 crc kubenswrapper[4923]: I1009 10:18:41.573177 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp4hl" event={"ID":"ecb6df4f-286a-4725-af44-c114db0a5e80","Type":"ContainerDied","Data":"14de722b2c88189f9520ff9e4472600ad926f97b972a95f0c750503f635ad9dc"} Oct 09 10:18:41 crc kubenswrapper[4923]: I1009 10:18:41.573216 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp4hl" event={"ID":"ecb6df4f-286a-4725-af44-c114db0a5e80","Type":"ContainerStarted","Data":"2aaaa23cd615ce4383c5754f239e557b90e98be8abe92927aa3cb9ab15d59f8f"} Oct 09 10:18:41 crc kubenswrapper[4923]: I1009 10:18:41.575501 4923 generic.go:334] "Generic (PLEG): container finished" podID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerID="dba38e9e96c4990c6ef817562e53fed35fc380ce87b25e6334c908f7d9d7ea6e" exitCode=0 Oct 09 10:18:41 crc kubenswrapper[4923]: I1009 10:18:41.575560 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5954" event={"ID":"17c9e5de-ac30-49e6-b6e4-46507b57b70d","Type":"ContainerDied","Data":"dba38e9e96c4990c6ef817562e53fed35fc380ce87b25e6334c908f7d9d7ea6e"} Oct 09 10:18:42 crc kubenswrapper[4923]: I1009 10:18:42.584161 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp4hl" event={"ID":"ecb6df4f-286a-4725-af44-c114db0a5e80","Type":"ContainerStarted","Data":"893de0841386d81307e177c80cc196b3a39f6ce7ee60aae79242379cb15249b9"} Oct 09 10:18:42 crc kubenswrapper[4923]: I1009 10:18:42.586998 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5954" event={"ID":"17c9e5de-ac30-49e6-b6e4-46507b57b70d","Type":"ContainerStarted","Data":"da99d7fa54c88a880e916394860d4ba6d05005267dc12403372faebb6d7f89b1"} Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.003541 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.020949 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q5954" podStartSLOduration=2.580560271 podStartE2EDuration="4.020911164s" podCreationTimestamp="2025-10-09 10:18:39 +0000 UTC" firstStartedPulling="2025-10-09 10:18:40.556142213 +0000 UTC m=+806.624323979" lastFinishedPulling="2025-10-09 10:18:41.996493116 +0000 UTC m=+808.064674872" observedRunningTime="2025-10-09 10:18:42.640840804 +0000 UTC m=+808.709022570" watchObservedRunningTime="2025-10-09 10:18:43.020911164 +0000 UTC m=+809.089092920" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.157243 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-util\") pod \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.157453 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-bundle\") pod \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.157549 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-flw9d\" (UniqueName: \"kubernetes.io/projected/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-kube-api-access-flw9d\") pod \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\" (UID: \"1a2f9a94-3b55-48ee-aa73-d22e8bde118c\") " Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.158258 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-bundle" (OuterVolumeSpecName: "bundle") pod "1a2f9a94-3b55-48ee-aa73-d22e8bde118c" (UID: "1a2f9a94-3b55-48ee-aa73-d22e8bde118c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.164825 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-kube-api-access-flw9d" (OuterVolumeSpecName: "kube-api-access-flw9d") pod "1a2f9a94-3b55-48ee-aa73-d22e8bde118c" (UID: "1a2f9a94-3b55-48ee-aa73-d22e8bde118c"). InnerVolumeSpecName "kube-api-access-flw9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.172195 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-util" (OuterVolumeSpecName: "util") pod "1a2f9a94-3b55-48ee-aa73-d22e8bde118c" (UID: "1a2f9a94-3b55-48ee-aa73-d22e8bde118c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.259705 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-flw9d\" (UniqueName: \"kubernetes.io/projected/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-kube-api-access-flw9d\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.259811 4923 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-util\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.259833 4923 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1a2f9a94-3b55-48ee-aa73-d22e8bde118c-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.605595 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.605646 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp" event={"ID":"1a2f9a94-3b55-48ee-aa73-d22e8bde118c","Type":"ContainerDied","Data":"ecc991c4a8a3b5d4bc45d25f1dcc0a66f1889af1784105744a832791ee02dfdf"} Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.605697 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ecc991c4a8a3b5d4bc45d25f1dcc0a66f1889af1784105744a832791ee02dfdf" Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.608049 4923 generic.go:334] "Generic (PLEG): container finished" podID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerID="893de0841386d81307e177c80cc196b3a39f6ce7ee60aae79242379cb15249b9" exitCode=0 Oct 09 10:18:43 crc kubenswrapper[4923]: I1009 10:18:43.608130 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp4hl" event={"ID":"ecb6df4f-286a-4725-af44-c114db0a5e80","Type":"ContainerDied","Data":"893de0841386d81307e177c80cc196b3a39f6ce7ee60aae79242379cb15249b9"} Oct 09 10:18:44 crc kubenswrapper[4923]: I1009 10:18:44.615940 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp4hl" event={"ID":"ecb6df4f-286a-4725-af44-c114db0a5e80","Type":"ContainerStarted","Data":"25ba53793e3777281779294b116a6d0ccdca68fca76e95439e792013156121c3"} Oct 09 10:18:44 crc kubenswrapper[4923]: I1009 10:18:44.638361 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wp4hl" podStartSLOduration=2.05199686 podStartE2EDuration="4.638345004s" podCreationTimestamp="2025-10-09 10:18:40 +0000 UTC" firstStartedPulling="2025-10-09 10:18:41.574932304 +0000 UTC m=+807.643114100" lastFinishedPulling="2025-10-09 10:18:44.161280488 +0000 UTC m=+810.229462244" observedRunningTime="2025-10-09 10:18:44.636876762 +0000 UTC m=+810.705058528" watchObservedRunningTime="2025-10-09 10:18:44.638345004 +0000 UTC m=+810.706526750" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.327418 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4"] Oct 09 10:18:48 crc kubenswrapper[4923]: E1009 10:18:48.328167 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerName="pull" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.328178 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerName="pull" Oct 09 10:18:48 crc kubenswrapper[4923]: E1009 10:18:48.328202 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerName="extract" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.328208 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerName="extract" Oct 09 10:18:48 crc kubenswrapper[4923]: E1009 10:18:48.328220 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerName="util" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.328226 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerName="util" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.328323 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a2f9a94-3b55-48ee-aa73-d22e8bde118c" containerName="extract" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.329029 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.331513 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-828hf" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.338506 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7x9n\" (UniqueName: \"kubernetes.io/projected/9aeb5354-727c-44fa-85dc-4921a593090f-kube-api-access-z7x9n\") pod \"openstack-operator-controller-operator-848c57cb5c-f88g4\" (UID: \"9aeb5354-727c-44fa-85dc-4921a593090f\") " pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.368211 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4"] Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.439641 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7x9n\" (UniqueName: \"kubernetes.io/projected/9aeb5354-727c-44fa-85dc-4921a593090f-kube-api-access-z7x9n\") pod \"openstack-operator-controller-operator-848c57cb5c-f88g4\" (UID: \"9aeb5354-727c-44fa-85dc-4921a593090f\") " pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.481874 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7x9n\" (UniqueName: \"kubernetes.io/projected/9aeb5354-727c-44fa-85dc-4921a593090f-kube-api-access-z7x9n\") pod \"openstack-operator-controller-operator-848c57cb5c-f88g4\" (UID: \"9aeb5354-727c-44fa-85dc-4921a593090f\") " pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" Oct 09 10:18:48 crc kubenswrapper[4923]: I1009 10:18:48.647429 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" Oct 09 10:18:49 crc kubenswrapper[4923]: I1009 10:18:49.132585 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4"] Oct 09 10:18:49 crc kubenswrapper[4923]: W1009 10:18:49.142249 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9aeb5354_727c_44fa_85dc_4921a593090f.slice/crio-085948d80608a446f03317e618ace39e81ba2c8fc0cac433c78f688a9cfa3763 WatchSource:0}: Error finding container 085948d80608a446f03317e618ace39e81ba2c8fc0cac433c78f688a9cfa3763: Status 404 returned error can't find the container with id 085948d80608a446f03317e618ace39e81ba2c8fc0cac433c78f688a9cfa3763 Oct 09 10:18:49 crc kubenswrapper[4923]: I1009 10:18:49.581809 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:49 crc kubenswrapper[4923]: I1009 10:18:49.581902 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:49 crc kubenswrapper[4923]: I1009 10:18:49.635162 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:49 crc kubenswrapper[4923]: I1009 10:18:49.657917 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" event={"ID":"9aeb5354-727c-44fa-85dc-4921a593090f","Type":"ContainerStarted","Data":"085948d80608a446f03317e618ace39e81ba2c8fc0cac433c78f688a9cfa3763"} Oct 09 10:18:49 crc kubenswrapper[4923]: I1009 10:18:49.708079 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:50 crc kubenswrapper[4923]: I1009 10:18:50.563256 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:50 crc kubenswrapper[4923]: I1009 10:18:50.563342 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:50 crc kubenswrapper[4923]: I1009 10:18:50.613644 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:50 crc kubenswrapper[4923]: I1009 10:18:50.719427 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:52 crc kubenswrapper[4923]: I1009 10:18:52.612361 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5954"] Oct 09 10:18:52 crc kubenswrapper[4923]: I1009 10:18:52.613086 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q5954" podUID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerName="registry-server" containerID="cri-o://da99d7fa54c88a880e916394860d4ba6d05005267dc12403372faebb6d7f89b1" gracePeriod=2 Oct 09 10:18:53 crc kubenswrapper[4923]: I1009 10:18:53.693332 4923 generic.go:334] "Generic (PLEG): container finished" podID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerID="da99d7fa54c88a880e916394860d4ba6d05005267dc12403372faebb6d7f89b1" exitCode=0 Oct 09 10:18:53 crc kubenswrapper[4923]: I1009 10:18:53.693381 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5954" event={"ID":"17c9e5de-ac30-49e6-b6e4-46507b57b70d","Type":"ContainerDied","Data":"da99d7fa54c88a880e916394860d4ba6d05005267dc12403372faebb6d7f89b1"} Oct 09 10:18:54 crc kubenswrapper[4923]: I1009 10:18:54.201197 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wp4hl"] Oct 09 10:18:54 crc kubenswrapper[4923]: I1009 10:18:54.201567 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wp4hl" podUID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerName="registry-server" containerID="cri-o://25ba53793e3777281779294b116a6d0ccdca68fca76e95439e792013156121c3" gracePeriod=2 Oct 09 10:18:54 crc kubenswrapper[4923]: I1009 10:18:54.707020 4923 generic.go:334] "Generic (PLEG): container finished" podID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerID="25ba53793e3777281779294b116a6d0ccdca68fca76e95439e792013156121c3" exitCode=0 Oct 09 10:18:54 crc kubenswrapper[4923]: I1009 10:18:54.707123 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp4hl" event={"ID":"ecb6df4f-286a-4725-af44-c114db0a5e80","Type":"ContainerDied","Data":"25ba53793e3777281779294b116a6d0ccdca68fca76e95439e792013156121c3"} Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.068069 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.080538 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.184478 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-utilities\") pod \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.184602 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpnbz\" (UniqueName: \"kubernetes.io/projected/17c9e5de-ac30-49e6-b6e4-46507b57b70d-kube-api-access-tpnbz\") pod \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.184661 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-catalog-content\") pod \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\" (UID: \"17c9e5de-ac30-49e6-b6e4-46507b57b70d\") " Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.184718 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drk4k\" (UniqueName: \"kubernetes.io/projected/ecb6df4f-286a-4725-af44-c114db0a5e80-kube-api-access-drk4k\") pod \"ecb6df4f-286a-4725-af44-c114db0a5e80\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.184772 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-catalog-content\") pod \"ecb6df4f-286a-4725-af44-c114db0a5e80\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.184902 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-utilities\") pod \"ecb6df4f-286a-4725-af44-c114db0a5e80\" (UID: \"ecb6df4f-286a-4725-af44-c114db0a5e80\") " Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.185945 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-utilities" (OuterVolumeSpecName: "utilities") pod "17c9e5de-ac30-49e6-b6e4-46507b57b70d" (UID: "17c9e5de-ac30-49e6-b6e4-46507b57b70d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.186235 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-utilities" (OuterVolumeSpecName: "utilities") pod "ecb6df4f-286a-4725-af44-c114db0a5e80" (UID: "ecb6df4f-286a-4725-af44-c114db0a5e80"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.191354 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecb6df4f-286a-4725-af44-c114db0a5e80-kube-api-access-drk4k" (OuterVolumeSpecName: "kube-api-access-drk4k") pod "ecb6df4f-286a-4725-af44-c114db0a5e80" (UID: "ecb6df4f-286a-4725-af44-c114db0a5e80"). InnerVolumeSpecName "kube-api-access-drk4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.195844 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17c9e5de-ac30-49e6-b6e4-46507b57b70d-kube-api-access-tpnbz" (OuterVolumeSpecName: "kube-api-access-tpnbz") pod "17c9e5de-ac30-49e6-b6e4-46507b57b70d" (UID: "17c9e5de-ac30-49e6-b6e4-46507b57b70d"). InnerVolumeSpecName "kube-api-access-tpnbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.203284 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17c9e5de-ac30-49e6-b6e4-46507b57b70d" (UID: "17c9e5de-ac30-49e6-b6e4-46507b57b70d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.265029 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ecb6df4f-286a-4725-af44-c114db0a5e80" (UID: "ecb6df4f-286a-4725-af44-c114db0a5e80"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.287076 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.287125 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ecb6df4f-286a-4725-af44-c114db0a5e80-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.287134 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.287146 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpnbz\" (UniqueName: \"kubernetes.io/projected/17c9e5de-ac30-49e6-b6e4-46507b57b70d-kube-api-access-tpnbz\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.287159 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17c9e5de-ac30-49e6-b6e4-46507b57b70d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.287171 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drk4k\" (UniqueName: \"kubernetes.io/projected/ecb6df4f-286a-4725-af44-c114db0a5e80-kube-api-access-drk4k\") on node \"crc\" DevicePath \"\"" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.719091 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wp4hl" event={"ID":"ecb6df4f-286a-4725-af44-c114db0a5e80","Type":"ContainerDied","Data":"2aaaa23cd615ce4383c5754f239e557b90e98be8abe92927aa3cb9ab15d59f8f"} Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.719157 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wp4hl" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.719207 4923 scope.go:117] "RemoveContainer" containerID="25ba53793e3777281779294b116a6d0ccdca68fca76e95439e792013156121c3" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.721675 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" event={"ID":"9aeb5354-727c-44fa-85dc-4921a593090f","Type":"ContainerStarted","Data":"ae749477ba39ea025dd6162b3445b514c0bed570c20bb974f7fd8a031e5efa5f"} Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.723303 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q5954" event={"ID":"17c9e5de-ac30-49e6-b6e4-46507b57b70d","Type":"ContainerDied","Data":"b4488e4491ea4a678d3812d7c4389f72f382c2704863adbe7aafe51ade35cbef"} Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.723404 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q5954" Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.762201 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5954"] Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.767439 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q5954"] Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.783932 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wp4hl"] Oct 09 10:18:55 crc kubenswrapper[4923]: I1009 10:18:55.789420 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wp4hl"] Oct 09 10:18:56 crc kubenswrapper[4923]: I1009 10:18:56.256893 4923 scope.go:117] "RemoveContainer" containerID="893de0841386d81307e177c80cc196b3a39f6ce7ee60aae79242379cb15249b9" Oct 09 10:18:56 crc kubenswrapper[4923]: I1009 10:18:56.386441 4923 scope.go:117] "RemoveContainer" containerID="14de722b2c88189f9520ff9e4472600ad926f97b972a95f0c750503f635ad9dc" Oct 09 10:18:56 crc kubenswrapper[4923]: I1009 10:18:56.534631 4923 scope.go:117] "RemoveContainer" containerID="da99d7fa54c88a880e916394860d4ba6d05005267dc12403372faebb6d7f89b1" Oct 09 10:18:56 crc kubenswrapper[4923]: I1009 10:18:56.611843 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" path="/var/lib/kubelet/pods/17c9e5de-ac30-49e6-b6e4-46507b57b70d/volumes" Oct 09 10:18:56 crc kubenswrapper[4923]: I1009 10:18:56.612697 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecb6df4f-286a-4725-af44-c114db0a5e80" path="/var/lib/kubelet/pods/ecb6df4f-286a-4725-af44-c114db0a5e80/volumes" Oct 09 10:18:57 crc kubenswrapper[4923]: I1009 10:18:57.241881 4923 scope.go:117] "RemoveContainer" containerID="dba38e9e96c4990c6ef817562e53fed35fc380ce87b25e6334c908f7d9d7ea6e" Oct 09 10:18:57 crc kubenswrapper[4923]: I1009 10:18:57.279519 4923 scope.go:117] "RemoveContainer" containerID="e1581c9d2d2d252d94d02767378f06ba73ac9f135d07fdf220b1d53c7e081a6c" Oct 09 10:18:57 crc kubenswrapper[4923]: I1009 10:18:57.742133 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" event={"ID":"9aeb5354-727c-44fa-85dc-4921a593090f","Type":"ContainerStarted","Data":"09f2086bcc2d20707e039c3760222d4f5bbd8c8fc2dc6c7d63e16e401e3f86f5"} Oct 09 10:18:57 crc kubenswrapper[4923]: I1009 10:18:57.742701 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" Oct 09 10:18:57 crc kubenswrapper[4923]: I1009 10:18:57.772380 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" podStartSLOduration=1.611394132 podStartE2EDuration="9.772352807s" podCreationTimestamp="2025-10-09 10:18:48 +0000 UTC" firstStartedPulling="2025-10-09 10:18:49.145274404 +0000 UTC m=+815.213456160" lastFinishedPulling="2025-10-09 10:18:57.306233079 +0000 UTC m=+823.374414835" observedRunningTime="2025-10-09 10:18:57.769083316 +0000 UTC m=+823.837265092" watchObservedRunningTime="2025-10-09 10:18:57.772352807 +0000 UTC m=+823.840534563" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.007938 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tdxvj"] Oct 09 10:19:00 crc kubenswrapper[4923]: E1009 10:19:00.008719 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerName="extract-utilities" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.008734 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerName="extract-utilities" Oct 09 10:19:00 crc kubenswrapper[4923]: E1009 10:19:00.008745 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerName="extract-utilities" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.008770 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerName="extract-utilities" Oct 09 10:19:00 crc kubenswrapper[4923]: E1009 10:19:00.008789 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerName="extract-content" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.008798 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerName="extract-content" Oct 09 10:19:00 crc kubenswrapper[4923]: E1009 10:19:00.008812 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerName="extract-content" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.008818 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerName="extract-content" Oct 09 10:19:00 crc kubenswrapper[4923]: E1009 10:19:00.008833 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerName="registry-server" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.008842 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerName="registry-server" Oct 09 10:19:00 crc kubenswrapper[4923]: E1009 10:19:00.008852 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerName="registry-server" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.008859 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerName="registry-server" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.008973 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb6df4f-286a-4725-af44-c114db0a5e80" containerName="registry-server" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.008987 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="17c9e5de-ac30-49e6-b6e4-46507b57b70d" containerName="registry-server" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.009959 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.022701 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdxvj"] Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.079597 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkfzt\" (UniqueName: \"kubernetes.io/projected/31c3a49d-7443-44eb-88b3-daa5932d711b-kube-api-access-bkfzt\") pod \"community-operators-tdxvj\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.079668 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-utilities\") pod \"community-operators-tdxvj\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.079724 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-catalog-content\") pod \"community-operators-tdxvj\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.181522 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-catalog-content\") pod \"community-operators-tdxvj\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.181649 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkfzt\" (UniqueName: \"kubernetes.io/projected/31c3a49d-7443-44eb-88b3-daa5932d711b-kube-api-access-bkfzt\") pod \"community-operators-tdxvj\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.181697 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-utilities\") pod \"community-operators-tdxvj\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.182121 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-catalog-content\") pod \"community-operators-tdxvj\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.182203 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-utilities\") pod \"community-operators-tdxvj\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.206499 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkfzt\" (UniqueName: \"kubernetes.io/projected/31c3a49d-7443-44eb-88b3-daa5932d711b-kube-api-access-bkfzt\") pod \"community-operators-tdxvj\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.328397 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:00 crc kubenswrapper[4923]: I1009 10:19:00.874346 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tdxvj"] Oct 09 10:19:00 crc kubenswrapper[4923]: W1009 10:19:00.890335 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31c3a49d_7443_44eb_88b3_daa5932d711b.slice/crio-46ce17907c4ed85ad2c670080b51363fe8f1f2b35114275391e68006d74fef07 WatchSource:0}: Error finding container 46ce17907c4ed85ad2c670080b51363fe8f1f2b35114275391e68006d74fef07: Status 404 returned error can't find the container with id 46ce17907c4ed85ad2c670080b51363fe8f1f2b35114275391e68006d74fef07 Oct 09 10:19:01 crc kubenswrapper[4923]: I1009 10:19:01.776141 4923 generic.go:334] "Generic (PLEG): container finished" podID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerID="4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f" exitCode=0 Oct 09 10:19:01 crc kubenswrapper[4923]: I1009 10:19:01.776246 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdxvj" event={"ID":"31c3a49d-7443-44eb-88b3-daa5932d711b","Type":"ContainerDied","Data":"4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f"} Oct 09 10:19:01 crc kubenswrapper[4923]: I1009 10:19:01.776497 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdxvj" event={"ID":"31c3a49d-7443-44eb-88b3-daa5932d711b","Type":"ContainerStarted","Data":"46ce17907c4ed85ad2c670080b51363fe8f1f2b35114275391e68006d74fef07"} Oct 09 10:19:03 crc kubenswrapper[4923]: I1009 10:19:03.790698 4923 generic.go:334] "Generic (PLEG): container finished" podID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerID="151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef" exitCode=0 Oct 09 10:19:03 crc kubenswrapper[4923]: I1009 10:19:03.790779 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdxvj" event={"ID":"31c3a49d-7443-44eb-88b3-daa5932d711b","Type":"ContainerDied","Data":"151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef"} Oct 09 10:19:04 crc kubenswrapper[4923]: I1009 10:19:04.801006 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdxvj" event={"ID":"31c3a49d-7443-44eb-88b3-daa5932d711b","Type":"ContainerStarted","Data":"f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2"} Oct 09 10:19:04 crc kubenswrapper[4923]: I1009 10:19:04.826047 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tdxvj" podStartSLOduration=3.2134080640000002 podStartE2EDuration="5.826021693s" podCreationTimestamp="2025-10-09 10:18:59 +0000 UTC" firstStartedPulling="2025-10-09 10:19:01.778028006 +0000 UTC m=+827.846209762" lastFinishedPulling="2025-10-09 10:19:04.390641625 +0000 UTC m=+830.458823391" observedRunningTime="2025-10-09 10:19:04.820094547 +0000 UTC m=+830.888276323" watchObservedRunningTime="2025-10-09 10:19:04.826021693 +0000 UTC m=+830.894203449" Oct 09 10:19:08 crc kubenswrapper[4923]: I1009 10:19:08.650922 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" Oct 09 10:19:10 crc kubenswrapper[4923]: I1009 10:19:10.329376 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:10 crc kubenswrapper[4923]: I1009 10:19:10.330016 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:10 crc kubenswrapper[4923]: I1009 10:19:10.376070 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:10 crc kubenswrapper[4923]: I1009 10:19:10.913203 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:11 crc kubenswrapper[4923]: I1009 10:19:11.414924 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdxvj"] Oct 09 10:19:12 crc kubenswrapper[4923]: I1009 10:19:12.853725 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tdxvj" podUID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerName="registry-server" containerID="cri-o://f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2" gracePeriod=2 Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.229510 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.295656 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bkfzt\" (UniqueName: \"kubernetes.io/projected/31c3a49d-7443-44eb-88b3-daa5932d711b-kube-api-access-bkfzt\") pod \"31c3a49d-7443-44eb-88b3-daa5932d711b\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.295773 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-catalog-content\") pod \"31c3a49d-7443-44eb-88b3-daa5932d711b\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.296013 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-utilities\") pod \"31c3a49d-7443-44eb-88b3-daa5932d711b\" (UID: \"31c3a49d-7443-44eb-88b3-daa5932d711b\") " Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.297125 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-utilities" (OuterVolumeSpecName: "utilities") pod "31c3a49d-7443-44eb-88b3-daa5932d711b" (UID: "31c3a49d-7443-44eb-88b3-daa5932d711b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.302084 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31c3a49d-7443-44eb-88b3-daa5932d711b-kube-api-access-bkfzt" (OuterVolumeSpecName: "kube-api-access-bkfzt") pod "31c3a49d-7443-44eb-88b3-daa5932d711b" (UID: "31c3a49d-7443-44eb-88b3-daa5932d711b"). InnerVolumeSpecName "kube-api-access-bkfzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.351331 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31c3a49d-7443-44eb-88b3-daa5932d711b" (UID: "31c3a49d-7443-44eb-88b3-daa5932d711b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.397494 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.397545 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bkfzt\" (UniqueName: \"kubernetes.io/projected/31c3a49d-7443-44eb-88b3-daa5932d711b-kube-api-access-bkfzt\") on node \"crc\" DevicePath \"\"" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.397561 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31c3a49d-7443-44eb-88b3-daa5932d711b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.866087 4923 generic.go:334] "Generic (PLEG): container finished" podID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerID="f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2" exitCode=0 Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.866144 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdxvj" event={"ID":"31c3a49d-7443-44eb-88b3-daa5932d711b","Type":"ContainerDied","Data":"f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2"} Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.866192 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tdxvj" event={"ID":"31c3a49d-7443-44eb-88b3-daa5932d711b","Type":"ContainerDied","Data":"46ce17907c4ed85ad2c670080b51363fe8f1f2b35114275391e68006d74fef07"} Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.866218 4923 scope.go:117] "RemoveContainer" containerID="f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.866220 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tdxvj" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.899792 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tdxvj"] Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.900047 4923 scope.go:117] "RemoveContainer" containerID="151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.908432 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tdxvj"] Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.925610 4923 scope.go:117] "RemoveContainer" containerID="4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.943865 4923 scope.go:117] "RemoveContainer" containerID="f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2" Oct 09 10:19:13 crc kubenswrapper[4923]: E1009 10:19:13.944399 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2\": container with ID starting with f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2 not found: ID does not exist" containerID="f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.944456 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2"} err="failed to get container status \"f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2\": rpc error: code = NotFound desc = could not find container \"f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2\": container with ID starting with f39c8fc5af9bcd865c554c105f9812ae7b6c8bab7742ca046fe4053c652fe1d2 not found: ID does not exist" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.944497 4923 scope.go:117] "RemoveContainer" containerID="151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef" Oct 09 10:19:13 crc kubenswrapper[4923]: E1009 10:19:13.945006 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef\": container with ID starting with 151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef not found: ID does not exist" containerID="151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.945038 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef"} err="failed to get container status \"151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef\": rpc error: code = NotFound desc = could not find container \"151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef\": container with ID starting with 151226b9459473a8746bd8c9788dc99e7025f1679340e942cb2965879f5168ef not found: ID does not exist" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.945055 4923 scope.go:117] "RemoveContainer" containerID="4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f" Oct 09 10:19:13 crc kubenswrapper[4923]: E1009 10:19:13.945347 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f\": container with ID starting with 4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f not found: ID does not exist" containerID="4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f" Oct 09 10:19:13 crc kubenswrapper[4923]: I1009 10:19:13.945384 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f"} err="failed to get container status \"4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f\": rpc error: code = NotFound desc = could not find container \"4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f\": container with ID starting with 4b8d2bb80b6e003d2ca97d0864e44e7212ec8e5ea4cf9b41de48bf5b510e9a3f not found: ID does not exist" Oct 09 10:19:14 crc kubenswrapper[4923]: I1009 10:19:14.621595 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31c3a49d-7443-44eb-88b3-daa5932d711b" path="/var/lib/kubelet/pods/31c3a49d-7443-44eb-88b3-daa5932d711b/volumes" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.072474 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t"] Oct 09 10:19:26 crc kubenswrapper[4923]: E1009 10:19:26.073737 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerName="extract-utilities" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.073775 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerName="extract-utilities" Oct 09 10:19:26 crc kubenswrapper[4923]: E1009 10:19:26.073801 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerName="extract-content" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.073812 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerName="extract-content" Oct 09 10:19:26 crc kubenswrapper[4923]: E1009 10:19:26.073841 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerName="registry-server" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.073849 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerName="registry-server" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.074009 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="31c3a49d-7443-44eb-88b3-daa5932d711b" containerName="registry-server" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.074909 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.078039 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-mvwx2" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.091673 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.097200 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.105656 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.124095 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-lrjqd" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.139830 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.155568 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.159634 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-4vbr6" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.173210 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.190631 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.206821 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.208902 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.214844 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.216086 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.217413 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-2g6w9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.218634 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-xfg7f" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.223618 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.225465 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqk84\" (UniqueName: \"kubernetes.io/projected/4466effc-4c4f-4023-94bc-d1e2f48d5d3b-kube-api-access-wqk84\") pod \"cinder-operator-controller-manager-7b7fb68549-lpvgj\" (UID: \"4466effc-4c4f-4023-94bc-d1e2f48d5d3b\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.225551 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d4nm\" (UniqueName: \"kubernetes.io/projected/91f871ca-9617-49c0-877f-ee0bcf938bee-kube-api-access-9d4nm\") pod \"barbican-operator-controller-manager-658bdf4b74-h294t\" (UID: \"91f871ca-9617-49c0-877f-ee0bcf938bee\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.230827 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.230907 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.232382 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.238776 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-j47c7" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.239057 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.252496 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.261226 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.266779 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-z6b7d" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.267396 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.282126 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.286717 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.307852 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.315163 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-bhssl" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.319144 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.320616 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.331842 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-v655m" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.333038 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vv8mr\" (UniqueName: \"kubernetes.io/projected/1438c7f5-06a5-49d0-be96-7fbee10b3010-kube-api-access-vv8mr\") pod \"glance-operator-controller-manager-84b9b84486-fhsk4\" (UID: \"1438c7f5-06a5-49d0-be96-7fbee10b3010\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.333101 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwglh\" (UniqueName: \"kubernetes.io/projected/36c97947-5f3f-4d0e-8125-0663f04f9eaf-kube-api-access-pwglh\") pod \"heat-operator-controller-manager-858f76bbdd-zjhgq\" (UID: \"36c97947-5f3f-4d0e-8125-0663f04f9eaf\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.333144 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqk84\" (UniqueName: \"kubernetes.io/projected/4466effc-4c4f-4023-94bc-d1e2f48d5d3b-kube-api-access-wqk84\") pod \"cinder-operator-controller-manager-7b7fb68549-lpvgj\" (UID: \"4466effc-4c4f-4023-94bc-d1e2f48d5d3b\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.333226 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d4nm\" (UniqueName: \"kubernetes.io/projected/91f871ca-9617-49c0-877f-ee0bcf938bee-kube-api-access-9d4nm\") pod \"barbican-operator-controller-manager-658bdf4b74-h294t\" (UID: \"91f871ca-9617-49c0-877f-ee0bcf938bee\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.333340 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8gr5\" (UniqueName: \"kubernetes.io/projected/4609ba20-418f-4134-b671-10154112f387-kube-api-access-v8gr5\") pod \"designate-operator-controller-manager-85d5d9dd78-lcs47\" (UID: \"4609ba20-418f-4134-b671-10154112f387\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.333521 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5chsz\" (UniqueName: \"kubernetes.io/projected/a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8-kube-api-access-5chsz\") pod \"horizon-operator-controller-manager-7ffbcb7588-gdprs\" (UID: \"a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.341966 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.377948 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.380001 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d4nm\" (UniqueName: \"kubernetes.io/projected/91f871ca-9617-49c0-877f-ee0bcf938bee-kube-api-access-9d4nm\") pod \"barbican-operator-controller-manager-658bdf4b74-h294t\" (UID: \"91f871ca-9617-49c0-877f-ee0bcf938bee\") " pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.383616 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.385162 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.386580 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqk84\" (UniqueName: \"kubernetes.io/projected/4466effc-4c4f-4023-94bc-d1e2f48d5d3b-kube-api-access-wqk84\") pod \"cinder-operator-controller-manager-7b7fb68549-lpvgj\" (UID: \"4466effc-4c4f-4023-94bc-d1e2f48d5d3b\") " pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.388275 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-h97gx" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.389659 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.390806 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.395706 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-xxpv8" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.400419 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.420101 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.436551 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5fba54a-5e97-419a-8950-65bd3908aadb-cert\") pod \"infra-operator-controller-manager-656bcbd775-td8lq\" (UID: \"c5fba54a-5e97-419a-8950-65bd3908aadb\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.436608 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8gr5\" (UniqueName: \"kubernetes.io/projected/4609ba20-418f-4134-b671-10154112f387-kube-api-access-v8gr5\") pod \"designate-operator-controller-manager-85d5d9dd78-lcs47\" (UID: \"4609ba20-418f-4134-b671-10154112f387\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.436663 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5chsz\" (UniqueName: \"kubernetes.io/projected/a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8-kube-api-access-5chsz\") pod \"horizon-operator-controller-manager-7ffbcb7588-gdprs\" (UID: \"a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.436688 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h66t6\" (UniqueName: \"kubernetes.io/projected/c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72-kube-api-access-h66t6\") pod \"ironic-operator-controller-manager-9c5c78d49-n8tdz\" (UID: \"c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.436716 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzfk6\" (UniqueName: \"kubernetes.io/projected/c5fba54a-5e97-419a-8950-65bd3908aadb-kube-api-access-hzfk6\") pod \"infra-operator-controller-manager-656bcbd775-td8lq\" (UID: \"c5fba54a-5e97-419a-8950-65bd3908aadb\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.436738 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vv8mr\" (UniqueName: \"kubernetes.io/projected/1438c7f5-06a5-49d0-be96-7fbee10b3010-kube-api-access-vv8mr\") pod \"glance-operator-controller-manager-84b9b84486-fhsk4\" (UID: \"1438c7f5-06a5-49d0-be96-7fbee10b3010\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.436780 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c94wq\" (UniqueName: \"kubernetes.io/projected/2ad61b43-c2ba-4dc9-8376-e75817b9e153-kube-api-access-c94wq\") pod \"keystone-operator-controller-manager-55b6b7c7b8-qrh4j\" (UID: \"2ad61b43-c2ba-4dc9-8376-e75817b9e153\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.436809 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwglh\" (UniqueName: \"kubernetes.io/projected/36c97947-5f3f-4d0e-8125-0663f04f9eaf-kube-api-access-pwglh\") pod \"heat-operator-controller-manager-858f76bbdd-zjhgq\" (UID: \"36c97947-5f3f-4d0e-8125-0663f04f9eaf\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.444195 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.453743 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.453920 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.465488 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-9rnpn" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.482278 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8gr5\" (UniqueName: \"kubernetes.io/projected/4609ba20-418f-4134-b671-10154112f387-kube-api-access-v8gr5\") pod \"designate-operator-controller-manager-85d5d9dd78-lcs47\" (UID: \"4609ba20-418f-4134-b671-10154112f387\") " pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.485667 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.490108 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.498380 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwglh\" (UniqueName: \"kubernetes.io/projected/36c97947-5f3f-4d0e-8125-0663f04f9eaf-kube-api-access-pwglh\") pod \"heat-operator-controller-manager-858f76bbdd-zjhgq\" (UID: \"36c97947-5f3f-4d0e-8125-0663f04f9eaf\") " pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.500643 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5chsz\" (UniqueName: \"kubernetes.io/projected/a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8-kube-api-access-5chsz\") pod \"horizon-operator-controller-manager-7ffbcb7588-gdprs\" (UID: \"a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8\") " pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.501616 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vv8mr\" (UniqueName: \"kubernetes.io/projected/1438c7f5-06a5-49d0-be96-7fbee10b3010-kube-api-access-vv8mr\") pod \"glance-operator-controller-manager-84b9b84486-fhsk4\" (UID: \"1438c7f5-06a5-49d0-be96-7fbee10b3010\") " pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.504129 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.510167 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.511396 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.518370 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-p2zlv" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.543831 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzfk6\" (UniqueName: \"kubernetes.io/projected/c5fba54a-5e97-419a-8950-65bd3908aadb-kube-api-access-hzfk6\") pod \"infra-operator-controller-manager-656bcbd775-td8lq\" (UID: \"c5fba54a-5e97-419a-8950-65bd3908aadb\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.543922 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c94wq\" (UniqueName: \"kubernetes.io/projected/2ad61b43-c2ba-4dc9-8376-e75817b9e153-kube-api-access-c94wq\") pod \"keystone-operator-controller-manager-55b6b7c7b8-qrh4j\" (UID: \"2ad61b43-c2ba-4dc9-8376-e75817b9e153\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.543960 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ntcl\" (UniqueName: \"kubernetes.io/projected/3fd48b51-1de2-4d9f-a790-55c95638fc2f-kube-api-access-9ntcl\") pod \"manila-operator-controller-manager-5f67fbc655-jz6s9\" (UID: \"3fd48b51-1de2-4d9f-a790-55c95638fc2f\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.543980 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxzds\" (UniqueName: \"kubernetes.io/projected/575ef7c8-b9ea-47ae-aa33-8c87596fff08-kube-api-access-qxzds\") pod \"mariadb-operator-controller-manager-f9fb45f8f-md5f2\" (UID: \"575ef7c8-b9ea-47ae-aa33-8c87596fff08\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.544044 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5fba54a-5e97-419a-8950-65bd3908aadb-cert\") pod \"infra-operator-controller-manager-656bcbd775-td8lq\" (UID: \"c5fba54a-5e97-419a-8950-65bd3908aadb\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.544096 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jbjh\" (UniqueName: \"kubernetes.io/projected/a44d21c5-6bf2-445e-9d72-2463571f1cf3-kube-api-access-8jbjh\") pod \"neutron-operator-controller-manager-79d585cb66-tp6xl\" (UID: \"a44d21c5-6bf2-445e-9d72-2463571f1cf3\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.544145 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h66t6\" (UniqueName: \"kubernetes.io/projected/c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72-kube-api-access-h66t6\") pod \"ironic-operator-controller-manager-9c5c78d49-n8tdz\" (UID: \"c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" Oct 09 10:19:26 crc kubenswrapper[4923]: E1009 10:19:26.545033 4923 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 09 10:19:26 crc kubenswrapper[4923]: E1009 10:19:26.545088 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c5fba54a-5e97-419a-8950-65bd3908aadb-cert podName:c5fba54a-5e97-419a-8950-65bd3908aadb nodeName:}" failed. No retries permitted until 2025-10-09 10:19:27.045066678 +0000 UTC m=+853.113248434 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c5fba54a-5e97-419a-8950-65bd3908aadb-cert") pod "infra-operator-controller-manager-656bcbd775-td8lq" (UID: "c5fba54a-5e97-419a-8950-65bd3908aadb") : secret "infra-operator-webhook-server-cert" not found Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.545272 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-55fjr"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.546545 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.552154 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.556159 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.557234 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.559619 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-67vjm" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.576533 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.604575 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-55fjr"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.688482 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h66t6\" (UniqueName: \"kubernetes.io/projected/c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72-kube-api-access-h66t6\") pod \"ironic-operator-controller-manager-9c5c78d49-n8tdz\" (UID: \"c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72\") " pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.724318 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c94wq\" (UniqueName: \"kubernetes.io/projected/2ad61b43-c2ba-4dc9-8376-e75817b9e153-kube-api-access-c94wq\") pod \"keystone-operator-controller-manager-55b6b7c7b8-qrh4j\" (UID: \"2ad61b43-c2ba-4dc9-8376-e75817b9e153\") " pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.742334 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jbjh\" (UniqueName: \"kubernetes.io/projected/a44d21c5-6bf2-445e-9d72-2463571f1cf3-kube-api-access-8jbjh\") pod \"neutron-operator-controller-manager-79d585cb66-tp6xl\" (UID: \"a44d21c5-6bf2-445e-9d72-2463571f1cf3\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.742393 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm2ms\" (UniqueName: \"kubernetes.io/projected/42a720da-4703-4d8b-9046-6c8e1643c880-kube-api-access-fm2ms\") pod \"nova-operator-controller-manager-5df598886f-55fjr\" (UID: \"42a720da-4703-4d8b-9046-6c8e1643c880\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.742614 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ntcl\" (UniqueName: \"kubernetes.io/projected/3fd48b51-1de2-4d9f-a790-55c95638fc2f-kube-api-access-9ntcl\") pod \"manila-operator-controller-manager-5f67fbc655-jz6s9\" (UID: \"3fd48b51-1de2-4d9f-a790-55c95638fc2f\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.742650 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxzds\" (UniqueName: \"kubernetes.io/projected/575ef7c8-b9ea-47ae-aa33-8c87596fff08-kube-api-access-qxzds\") pod \"mariadb-operator-controller-manager-f9fb45f8f-md5f2\" (UID: \"575ef7c8-b9ea-47ae-aa33-8c87596fff08\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.742812 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrpqx\" (UniqueName: \"kubernetes.io/projected/120ce970-aff3-40cb-8724-957a034fb914-kube-api-access-nrpqx\") pod \"octavia-operator-controller-manager-69fdcfc5f5-pbjk9\" (UID: \"120ce970-aff3-40cb-8724-957a034fb914\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.759742 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.770787 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.780586 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzfk6\" (UniqueName: \"kubernetes.io/projected/c5fba54a-5e97-419a-8950-65bd3908aadb-kube-api-access-hzfk6\") pod \"infra-operator-controller-manager-656bcbd775-td8lq\" (UID: \"c5fba54a-5e97-419a-8950-65bd3908aadb\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.785815 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-b7n4l" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.798939 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jbjh\" (UniqueName: \"kubernetes.io/projected/a44d21c5-6bf2-445e-9d72-2463571f1cf3-kube-api-access-8jbjh\") pod \"neutron-operator-controller-manager-79d585cb66-tp6xl\" (UID: \"a44d21c5-6bf2-445e-9d72-2463571f1cf3\") " pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.799747 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ntcl\" (UniqueName: \"kubernetes.io/projected/3fd48b51-1de2-4d9f-a790-55c95638fc2f-kube-api-access-9ntcl\") pod \"manila-operator-controller-manager-5f67fbc655-jz6s9\" (UID: \"3fd48b51-1de2-4d9f-a790-55c95638fc2f\") " pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.823074 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxzds\" (UniqueName: \"kubernetes.io/projected/575ef7c8-b9ea-47ae-aa33-8c87596fff08-kube-api-access-qxzds\") pod \"mariadb-operator-controller-manager-f9fb45f8f-md5f2\" (UID: \"575ef7c8-b9ea-47ae-aa33-8c87596fff08\") " pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.824905 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.827043 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.829728 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-w6gjk" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.830062 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.845572 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrpqx\" (UniqueName: \"kubernetes.io/projected/120ce970-aff3-40cb-8724-957a034fb914-kube-api-access-nrpqx\") pod \"octavia-operator-controller-manager-69fdcfc5f5-pbjk9\" (UID: \"120ce970-aff3-40cb-8724-957a034fb914\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.845991 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm2ms\" (UniqueName: \"kubernetes.io/projected/42a720da-4703-4d8b-9046-6c8e1643c880-kube-api-access-fm2ms\") pod \"nova-operator-controller-manager-5df598886f-55fjr\" (UID: \"42a720da-4703-4d8b-9046-6c8e1643c880\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.862246 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.882597 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm2ms\" (UniqueName: \"kubernetes.io/projected/42a720da-4703-4d8b-9046-6c8e1643c880-kube-api-access-fm2ms\") pod \"nova-operator-controller-manager-5df598886f-55fjr\" (UID: \"42a720da-4703-4d8b-9046-6c8e1643c880\") " pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.882603 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrpqx\" (UniqueName: \"kubernetes.io/projected/120ce970-aff3-40cb-8724-957a034fb914-kube-api-access-nrpqx\") pod \"octavia-operator-controller-manager-69fdcfc5f5-pbjk9\" (UID: \"120ce970-aff3-40cb-8724-957a034fb914\") " pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.891821 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.894318 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.897549 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-476hj" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.902404 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.918880 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.951787 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.957359 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs4tp\" (UniqueName: \"kubernetes.io/projected/6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed-kube-api-access-rs4tp\") pod \"ovn-operator-controller-manager-79db49b9fb-f89mg\" (UID: \"6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.957514 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98sfh\" (UniqueName: \"kubernetes.io/projected/29e93b24-6e22-4ade-9fd9-1742186d21fd-kube-api-access-98sfh\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bffmcv74\" (UID: \"29e93b24-6e22-4ade-9fd9-1742186d21fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.957774 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29e93b24-6e22-4ade-9fd9-1742186d21fd-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bffmcv74\" (UID: \"29e93b24-6e22-4ade-9fd9-1742186d21fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.961806 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.991871 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6"] Oct 09 10:19:26 crc kubenswrapper[4923]: I1009 10:19:26.994550 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.002626 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.005006 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wktgl" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.020120 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.020839 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.036900 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.045810 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.053194 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.060074 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs4tp\" (UniqueName: \"kubernetes.io/projected/6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed-kube-api-access-rs4tp\") pod \"ovn-operator-controller-manager-79db49b9fb-f89mg\" (UID: \"6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.060152 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d9g8\" (UniqueName: \"kubernetes.io/projected/ada28834-1248-46cb-b482-4a2d8ee77520-kube-api-access-5d9g8\") pod \"placement-operator-controller-manager-68b6c87b68-2cvzn\" (UID: \"ada28834-1248-46cb-b482-4a2d8ee77520\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.060224 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98sfh\" (UniqueName: \"kubernetes.io/projected/29e93b24-6e22-4ade-9fd9-1742186d21fd-kube-api-access-98sfh\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bffmcv74\" (UID: \"29e93b24-6e22-4ade-9fd9-1742186d21fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.060256 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29e93b24-6e22-4ade-9fd9-1742186d21fd-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bffmcv74\" (UID: \"29e93b24-6e22-4ade-9fd9-1742186d21fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.060340 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5fba54a-5e97-419a-8950-65bd3908aadb-cert\") pod \"infra-operator-controller-manager-656bcbd775-td8lq\" (UID: \"c5fba54a-5e97-419a-8950-65bd3908aadb\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.066145 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6"] Oct 09 10:19:27 crc kubenswrapper[4923]: E1009 10:19:27.066679 4923 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 09 10:19:27 crc kubenswrapper[4923]: E1009 10:19:27.066789 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/29e93b24-6e22-4ade-9fd9-1742186d21fd-cert podName:29e93b24-6e22-4ade-9fd9-1742186d21fd nodeName:}" failed. No retries permitted until 2025-10-09 10:19:27.566741529 +0000 UTC m=+853.634923285 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/29e93b24-6e22-4ade-9fd9-1742186d21fd-cert") pod "openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" (UID: "29e93b24-6e22-4ade-9fd9-1742186d21fd") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.066799 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c5fba54a-5e97-419a-8950-65bd3908aadb-cert\") pod \"infra-operator-controller-manager-656bcbd775-td8lq\" (UID: \"c5fba54a-5e97-419a-8950-65bd3908aadb\") " pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.076244 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.088642 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ht4ns" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.128911 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.130008 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-flh7l"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.131260 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.133678 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-4wp45" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.134036 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs4tp\" (UniqueName: \"kubernetes.io/projected/6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed-kube-api-access-rs4tp\") pod \"ovn-operator-controller-manager-79db49b9fb-f89mg\" (UID: \"6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed\") " pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.138097 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.139572 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.145366 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-xhmtv" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.148946 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-flh7l"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.150701 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98sfh\" (UniqueName: \"kubernetes.io/projected/29e93b24-6e22-4ade-9fd9-1742186d21fd-kube-api-access-98sfh\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bffmcv74\" (UID: \"29e93b24-6e22-4ade-9fd9-1742186d21fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.166492 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvh5k\" (UniqueName: \"kubernetes.io/projected/c5e2cc1d-7b01-4316-a960-5c47a420d6fd-kube-api-access-nvh5k\") pod \"swift-operator-controller-manager-db6d7f97b-xzpv6\" (UID: \"c5e2cc1d-7b01-4316-a960-5c47a420d6fd\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.166596 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d9g8\" (UniqueName: \"kubernetes.io/projected/ada28834-1248-46cb-b482-4a2d8ee77520-kube-api-access-5d9g8\") pod \"placement-operator-controller-manager-68b6c87b68-2cvzn\" (UID: \"ada28834-1248-46cb-b482-4a2d8ee77520\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.174270 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.191710 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.251521 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d9g8\" (UniqueName: \"kubernetes.io/projected/ada28834-1248-46cb-b482-4a2d8ee77520-kube-api-access-5d9g8\") pod \"placement-operator-controller-manager-68b6c87b68-2cvzn\" (UID: \"ada28834-1248-46cb-b482-4a2d8ee77520\") " pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.253220 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.254586 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.263770 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.264886 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-9jdrm" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.269914 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.272699 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrwvc\" (UniqueName: \"kubernetes.io/projected/f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8-kube-api-access-lrwvc\") pod \"test-operator-controller-manager-5458f77c4-flh7l\" (UID: \"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.272845 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qv69\" (UniqueName: \"kubernetes.io/projected/c0d0a77b-c472-4690-b169-becf1ea318dc-kube-api-access-4qv69\") pod \"watcher-operator-controller-manager-7f554bff7b-54cbp\" (UID: \"c0d0a77b-c472-4690-b169-becf1ea318dc\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.272909 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvh5k\" (UniqueName: \"kubernetes.io/projected/c5e2cc1d-7b01-4316-a960-5c47a420d6fd-kube-api-access-nvh5k\") pod \"swift-operator-controller-manager-db6d7f97b-xzpv6\" (UID: \"c5e2cc1d-7b01-4316-a960-5c47a420d6fd\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.272977 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fs9ch\" (UniqueName: \"kubernetes.io/projected/9e8c5a78-2a63-4219-8293-ea6d866a2d1b-kube-api-access-fs9ch\") pod \"telemetry-operator-controller-manager-67cfc6749b-hw8r6\" (UID: \"9e8c5a78-2a63-4219-8293-ea6d866a2d1b\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.316863 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.318441 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.324240 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-z2fks" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.332584 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.348669 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvh5k\" (UniqueName: \"kubernetes.io/projected/c5e2cc1d-7b01-4316-a960-5c47a420d6fd-kube-api-access-nvh5k\") pod \"swift-operator-controller-manager-db6d7f97b-xzpv6\" (UID: \"c5e2cc1d-7b01-4316-a960-5c47a420d6fd\") " pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.373942 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fs9ch\" (UniqueName: \"kubernetes.io/projected/9e8c5a78-2a63-4219-8293-ea6d866a2d1b-kube-api-access-fs9ch\") pod \"telemetry-operator-controller-manager-67cfc6749b-hw8r6\" (UID: \"9e8c5a78-2a63-4219-8293-ea6d866a2d1b\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.374042 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrwvc\" (UniqueName: \"kubernetes.io/projected/f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8-kube-api-access-lrwvc\") pod \"test-operator-controller-manager-5458f77c4-flh7l\" (UID: \"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.374092 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhhmz\" (UniqueName: \"kubernetes.io/projected/2ac22a8a-f293-4c68-85be-286bc263e77b-kube-api-access-fhhmz\") pod \"openstack-operator-controller-manager-7d6957655c-h7b88\" (UID: \"2ac22a8a-f293-4c68-85be-286bc263e77b\") " pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.374174 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qv69\" (UniqueName: \"kubernetes.io/projected/c0d0a77b-c472-4690-b169-becf1ea318dc-kube-api-access-4qv69\") pod \"watcher-operator-controller-manager-7f554bff7b-54cbp\" (UID: \"c0d0a77b-c472-4690-b169-becf1ea318dc\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.374203 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ac22a8a-f293-4c68-85be-286bc263e77b-cert\") pod \"openstack-operator-controller-manager-7d6957655c-h7b88\" (UID: \"2ac22a8a-f293-4c68-85be-286bc263e77b\") " pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.416488 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrwvc\" (UniqueName: \"kubernetes.io/projected/f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8-kube-api-access-lrwvc\") pod \"test-operator-controller-manager-5458f77c4-flh7l\" (UID: \"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8\") " pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.434323 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.452343 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qv69\" (UniqueName: \"kubernetes.io/projected/c0d0a77b-c472-4690-b169-becf1ea318dc-kube-api-access-4qv69\") pod \"watcher-operator-controller-manager-7f554bff7b-54cbp\" (UID: \"c0d0a77b-c472-4690-b169-becf1ea318dc\") " pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.452391 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fs9ch\" (UniqueName: \"kubernetes.io/projected/9e8c5a78-2a63-4219-8293-ea6d866a2d1b-kube-api-access-fs9ch\") pod \"telemetry-operator-controller-manager-67cfc6749b-hw8r6\" (UID: \"9e8c5a78-2a63-4219-8293-ea6d866a2d1b\") " pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.477448 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhhmz\" (UniqueName: \"kubernetes.io/projected/2ac22a8a-f293-4c68-85be-286bc263e77b-kube-api-access-fhhmz\") pod \"openstack-operator-controller-manager-7d6957655c-h7b88\" (UID: \"2ac22a8a-f293-4c68-85be-286bc263e77b\") " pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.477570 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mz7h\" (UniqueName: \"kubernetes.io/projected/cd75b71e-8663-4e84-a99f-2d902135917d-kube-api-access-9mz7h\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t\" (UID: \"cd75b71e-8663-4e84-a99f-2d902135917d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.477633 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ac22a8a-f293-4c68-85be-286bc263e77b-cert\") pod \"openstack-operator-controller-manager-7d6957655c-h7b88\" (UID: \"2ac22a8a-f293-4c68-85be-286bc263e77b\") " pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:27 crc kubenswrapper[4923]: E1009 10:19:27.478205 4923 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 09 10:19:27 crc kubenswrapper[4923]: E1009 10:19:27.478394 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2ac22a8a-f293-4c68-85be-286bc263e77b-cert podName:2ac22a8a-f293-4c68-85be-286bc263e77b nodeName:}" failed. No retries permitted until 2025-10-09 10:19:27.97832174 +0000 UTC m=+854.046503496 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2ac22a8a-f293-4c68-85be-286bc263e77b-cert") pod "openstack-operator-controller-manager-7d6957655c-h7b88" (UID: "2ac22a8a-f293-4c68-85be-286bc263e77b") : secret "webhook-server-cert" not found Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.487613 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t"] Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.502352 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhhmz\" (UniqueName: \"kubernetes.io/projected/2ac22a8a-f293-4c68-85be-286bc263e77b-kube-api-access-fhhmz\") pod \"openstack-operator-controller-manager-7d6957655c-h7b88\" (UID: \"2ac22a8a-f293-4c68-85be-286bc263e77b\") " pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.527075 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.579466 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mz7h\" (UniqueName: \"kubernetes.io/projected/cd75b71e-8663-4e84-a99f-2d902135917d-kube-api-access-9mz7h\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t\" (UID: \"cd75b71e-8663-4e84-a99f-2d902135917d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.579543 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29e93b24-6e22-4ade-9fd9-1742186d21fd-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bffmcv74\" (UID: \"29e93b24-6e22-4ade-9fd9-1742186d21fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.587316 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/29e93b24-6e22-4ade-9fd9-1742186d21fd-cert\") pod \"openstack-baremetal-operator-controller-manager-677c5f5bffmcv74\" (UID: \"29e93b24-6e22-4ade-9fd9-1742186d21fd\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.612805 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mz7h\" (UniqueName: \"kubernetes.io/projected/cd75b71e-8663-4e84-a99f-2d902135917d-kube-api-access-9mz7h\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t\" (UID: \"cd75b71e-8663-4e84-a99f-2d902135917d\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.628463 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.696059 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.698464 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.727471 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.767179 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:19:27 crc kubenswrapper[4923]: I1009 10:19:27.840719 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.014846 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ac22a8a-f293-4c68-85be-286bc263e77b-cert\") pod \"openstack-operator-controller-manager-7d6957655c-h7b88\" (UID: \"2ac22a8a-f293-4c68-85be-286bc263e77b\") " pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.031341 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" event={"ID":"91f871ca-9617-49c0-877f-ee0bcf938bee","Type":"ContainerStarted","Data":"3da173a8c4d1e8273af79d70ce1204c66e46b2c8cbeb8dfa0ed3afef3bb14917"} Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.040317 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2ac22a8a-f293-4c68-85be-286bc263e77b-cert\") pod \"openstack-operator-controller-manager-7d6957655c-h7b88\" (UID: \"2ac22a8a-f293-4c68-85be-286bc263e77b\") " pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.057279 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.076487 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47"] Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.088572 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4"] Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.096248 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj"] Oct 09 10:19:28 crc kubenswrapper[4923]: W1009 10:19:28.109018 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1438c7f5_06a5_49d0_be96_7fbee10b3010.slice/crio-32c9ebe846fe9ebb2423bea7e0ec691bd0a6b6598b35e994129fea13684aebea WatchSource:0}: Error finding container 32c9ebe846fe9ebb2423bea7e0ec691bd0a6b6598b35e994129fea13684aebea: Status 404 returned error can't find the container with id 32c9ebe846fe9ebb2423bea7e0ec691bd0a6b6598b35e994129fea13684aebea Oct 09 10:19:28 crc kubenswrapper[4923]: W1009 10:19:28.140575 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4609ba20_418f_4134_b671_10154112f387.slice/crio-d35f09f21e5a0b07172c31f9420b4da44bdbde5efacd8dd413a84823cd746a59 WatchSource:0}: Error finding container d35f09f21e5a0b07172c31f9420b4da44bdbde5efacd8dd413a84823cd746a59: Status 404 returned error can't find the container with id d35f09f21e5a0b07172c31f9420b4da44bdbde5efacd8dd413a84823cd746a59 Oct 09 10:19:28 crc kubenswrapper[4923]: W1009 10:19:28.151936 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4466effc_4c4f_4023_94bc_d1e2f48d5d3b.slice/crio-4921bcf4042ccc34afd34633f1968e37b621075a17d8a48138e61b97ef8ebda3 WatchSource:0}: Error finding container 4921bcf4042ccc34afd34633f1968e37b621075a17d8a48138e61b97ef8ebda3: Status 404 returned error can't find the container with id 4921bcf4042ccc34afd34633f1968e37b621075a17d8a48138e61b97ef8ebda3 Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.580585 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz"] Oct 09 10:19:28 crc kubenswrapper[4923]: W1009 10:19:28.585410 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7ebbe8d_b85b_479c_9fd8_e7ad2197bc72.slice/crio-85f10f1c922851d9848340bbe786f45015a0d9f666850bea558d6f087c954fbb WatchSource:0}: Error finding container 85f10f1c922851d9848340bbe786f45015a0d9f666850bea558d6f087c954fbb: Status 404 returned error can't find the container with id 85f10f1c922851d9848340bbe786f45015a0d9f666850bea558d6f087c954fbb Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.632042 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq"] Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.639237 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j"] Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.644369 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs"] Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.656677 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9"] Oct 09 10:19:28 crc kubenswrapper[4923]: W1009 10:19:28.671902 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36c97947_5f3f_4d0e_8125_0663f04f9eaf.slice/crio-eacd12aadacc7e9c255eca78f2d137e1f9af3ed205bfab8871b622d655a7860c WatchSource:0}: Error finding container eacd12aadacc7e9c255eca78f2d137e1f9af3ed205bfab8871b622d655a7860c: Status 404 returned error can't find the container with id eacd12aadacc7e9c255eca78f2d137e1f9af3ed205bfab8871b622d655a7860c Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.692423 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2"] Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.705993 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq"] Oct 09 10:19:28 crc kubenswrapper[4923]: W1009 10:19:28.712065 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fd48b51_1de2_4d9f_a790_55c95638fc2f.slice/crio-4d3e8103e3f741222a1fcfc83b2ff6087185791df9752a878b6042c640cb59a6 WatchSource:0}: Error finding container 4d3e8103e3f741222a1fcfc83b2ff6087185791df9752a878b6042c640cb59a6: Status 404 returned error can't find the container with id 4d3e8103e3f741222a1fcfc83b2ff6087185791df9752a878b6042c640cb59a6 Oct 09 10:19:28 crc kubenswrapper[4923]: I1009 10:19:28.726125 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5df598886f-55fjr"] Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.044139 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" event={"ID":"575ef7c8-b9ea-47ae-aa33-8c87596fff08","Type":"ContainerStarted","Data":"2c2a169e27c67947ffd2e55381a9f02fb8749296ab756108aea0615eded7a5af"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.045522 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" event={"ID":"42a720da-4703-4d8b-9046-6c8e1643c880","Type":"ContainerStarted","Data":"5ab5dc365044ab8f69072bc8aa0f08d4bfdb628f1e845c95a86b940225955fa5"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.047487 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" event={"ID":"36c97947-5f3f-4d0e-8125-0663f04f9eaf","Type":"ContainerStarted","Data":"eacd12aadacc7e9c255eca78f2d137e1f9af3ed205bfab8871b622d655a7860c"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.049175 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" event={"ID":"c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72","Type":"ContainerStarted","Data":"85f10f1c922851d9848340bbe786f45015a0d9f666850bea558d6f087c954fbb"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.050728 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" event={"ID":"1438c7f5-06a5-49d0-be96-7fbee10b3010","Type":"ContainerStarted","Data":"32c9ebe846fe9ebb2423bea7e0ec691bd0a6b6598b35e994129fea13684aebea"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.052628 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" event={"ID":"2ad61b43-c2ba-4dc9-8376-e75817b9e153","Type":"ContainerStarted","Data":"76975a7ed6f166b67c99b5b9f00c3569c38fe855efe4b2586e94d5074584cdd1"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.054621 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" event={"ID":"3fd48b51-1de2-4d9f-a790-55c95638fc2f","Type":"ContainerStarted","Data":"4d3e8103e3f741222a1fcfc83b2ff6087185791df9752a878b6042c640cb59a6"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.064392 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" event={"ID":"4466effc-4c4f-4023-94bc-d1e2f48d5d3b","Type":"ContainerStarted","Data":"4921bcf4042ccc34afd34633f1968e37b621075a17d8a48138e61b97ef8ebda3"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.071197 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" event={"ID":"c5fba54a-5e97-419a-8950-65bd3908aadb","Type":"ContainerStarted","Data":"42847ead3867dc1a747b496e1032811f59e126ef6f8a16c7fdfcef5ed7d04288"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.078844 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" event={"ID":"a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8","Type":"ContainerStarted","Data":"47c58df3aa34ca928ee167745d969f090283b73300742a3e1b0965e0e0380be9"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.080190 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" event={"ID":"4609ba20-418f-4134-b671-10154112f387","Type":"ContainerStarted","Data":"d35f09f21e5a0b07172c31f9420b4da44bdbde5efacd8dd413a84823cd746a59"} Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.085797 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74"] Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.098332 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-flh7l"] Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.108966 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn"] Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.244015 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88"] Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.255021 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg"] Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.280244 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp"] Oct 09 10:19:29 crc kubenswrapper[4923]: W1009 10:19:29.280953 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d7ad3b7_04a4_41d0_90fe_15f32fa9c0ed.slice/crio-fb74b72946951b7f9155fc1fcfcf55ee5c19994e9d972008c09050e8312ff6ea WatchSource:0}: Error finding container fb74b72946951b7f9155fc1fcfcf55ee5c19994e9d972008c09050e8312ff6ea: Status 404 returned error can't find the container with id fb74b72946951b7f9155fc1fcfcf55ee5c19994e9d972008c09050e8312ff6ea Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.290744 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6"] Oct 09 10:19:29 crc kubenswrapper[4923]: W1009 10:19:29.294704 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5e2cc1d_7b01_4316_a960_5c47a420d6fd.slice/crio-50cc43924fa76323f210a15e90fd72d1f0d968be448fc9383aa140aa4576b1aa WatchSource:0}: Error finding container 50cc43924fa76323f210a15e90fd72d1f0d968be448fc9383aa140aa4576b1aa: Status 404 returned error can't find the container with id 50cc43924fa76323f210a15e90fd72d1f0d968be448fc9383aa140aa4576b1aa Oct 09 10:19:29 crc kubenswrapper[4923]: W1009 10:19:29.297361 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0d0a77b_c472_4690_b169_becf1ea318dc.slice/crio-69ab7e024fc2b108b9b50b9e57715f8aa46f74dd0d2e7ebc2b0d35999aca0b70 WatchSource:0}: Error finding container 69ab7e024fc2b108b9b50b9e57715f8aa46f74dd0d2e7ebc2b0d35999aca0b70: Status 404 returned error can't find the container with id 69ab7e024fc2b108b9b50b9e57715f8aa46f74dd0d2e7ebc2b0d35999aca0b70 Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.302724 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6"] Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.308614 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl"] Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.313388 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4qv69,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-7f554bff7b-54cbp_openstack-operators(c0d0a77b-c472-4690-b169-becf1ea318dc): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.313590 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nvh5k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-db6d7f97b-xzpv6_openstack-operators(c5e2cc1d-7b01-4316-a960-5c47a420d6fd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.313946 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9"] Oct 09 10:19:29 crc kubenswrapper[4923]: I1009 10:19:29.316591 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t"] Oct 09 10:19:29 crc kubenswrapper[4923]: W1009 10:19:29.332231 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9e8c5a78_2a63_4219_8293_ea6d866a2d1b.slice/crio-6c031a3026abf6ee681b4e0259056bb74bef30ac25193a8bee087f9a909ee283 WatchSource:0}: Error finding container 6c031a3026abf6ee681b4e0259056bb74bef30ac25193a8bee087f9a909ee283: Status 404 returned error can't find the container with id 6c031a3026abf6ee681b4e0259056bb74bef30ac25193a8bee087f9a909ee283 Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.344862 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fs9ch,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-67cfc6749b-hw8r6_openstack-operators(9e8c5a78-2a63-4219-8293-ea6d866a2d1b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.379793 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9mz7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t_openstack-operators(cd75b71e-8663-4e84-a99f-2d902135917d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.379960 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8jbjh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-79d585cb66-tp6xl_openstack-operators(a44d21c5-6bf2-445e-9d72-2463571f1cf3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.380992 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" podUID="cd75b71e-8663-4e84-a99f-2d902135917d" Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.867300 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" podUID="c0d0a77b-c472-4690-b169-becf1ea318dc" Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.895775 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" podUID="9e8c5a78-2a63-4219-8293-ea6d866a2d1b" Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.896520 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" podUID="c5e2cc1d-7b01-4316-a960-5c47a420d6fd" Oct 09 10:19:29 crc kubenswrapper[4923]: E1009 10:19:29.910187 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" podUID="a44d21c5-6bf2-445e-9d72-2463571f1cf3" Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.136257 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" event={"ID":"120ce970-aff3-40cb-8724-957a034fb914","Type":"ContainerStarted","Data":"8dd886eb40bd2d84705a3a5be43a92fcdefb6c53a4d82e38b133ea46e3bce146"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.139578 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" event={"ID":"2ac22a8a-f293-4c68-85be-286bc263e77b","Type":"ContainerStarted","Data":"7ff2f1caaecc946af331eb1bd3d266160e7ea5160146876804fb77eaadc3a01a"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.139632 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" event={"ID":"2ac22a8a-f293-4c68-85be-286bc263e77b","Type":"ContainerStarted","Data":"be49a24fcbc17409c678c008b916fdb0da763b8d5c2efbddaa2e5995f984cdd8"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.142667 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" event={"ID":"9e8c5a78-2a63-4219-8293-ea6d866a2d1b","Type":"ContainerStarted","Data":"a91de193b7dfd18265eacacc4f0f9522a221e9ea78aaab35187e65e30245e579"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.142819 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" event={"ID":"9e8c5a78-2a63-4219-8293-ea6d866a2d1b","Type":"ContainerStarted","Data":"6c031a3026abf6ee681b4e0259056bb74bef30ac25193a8bee087f9a909ee283"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.152710 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" event={"ID":"cd75b71e-8663-4e84-a99f-2d902135917d","Type":"ContainerStarted","Data":"2e9db311546f82c694c79145b2dd14812732550e31c501d95758e9aa778f5736"} Oct 09 10:19:30 crc kubenswrapper[4923]: E1009 10:19:30.158096 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" podUID="9e8c5a78-2a63-4219-8293-ea6d866a2d1b" Oct 09 10:19:30 crc kubenswrapper[4923]: E1009 10:19:30.158211 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" podUID="cd75b71e-8663-4e84-a99f-2d902135917d" Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.161112 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" event={"ID":"c0d0a77b-c472-4690-b169-becf1ea318dc","Type":"ContainerStarted","Data":"870b38d4b8b4f751b0f8d2472385d577f45183967dd763c16ff8e62b13df55e2"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.161162 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" event={"ID":"c0d0a77b-c472-4690-b169-becf1ea318dc","Type":"ContainerStarted","Data":"69ab7e024fc2b108b9b50b9e57715f8aa46f74dd0d2e7ebc2b0d35999aca0b70"} Oct 09 10:19:30 crc kubenswrapper[4923]: E1009 10:19:30.185205 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" podUID="c0d0a77b-c472-4690-b169-becf1ea318dc" Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.195297 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" event={"ID":"a44d21c5-6bf2-445e-9d72-2463571f1cf3","Type":"ContainerStarted","Data":"be21f1a9c27fdfffc7343d6facaf5ab66cf6fc6d8aea48b147dbd32b8074f9d6"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.195387 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" event={"ID":"a44d21c5-6bf2-445e-9d72-2463571f1cf3","Type":"ContainerStarted","Data":"067ec450e1fbd91a3508d028c531c15f5398f76923a4bfe3bf8cb477128eacb9"} Oct 09 10:19:30 crc kubenswrapper[4923]: E1009 10:19:30.199455 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" podUID="a44d21c5-6bf2-445e-9d72-2463571f1cf3" Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.216011 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" event={"ID":"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8","Type":"ContainerStarted","Data":"d77859ab34b9a08bbc22b0e8b5719ecbd7bca4fe1b987fe715a3969059a3e909"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.223267 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" event={"ID":"29e93b24-6e22-4ade-9fd9-1742186d21fd","Type":"ContainerStarted","Data":"7da191c1ef2ffea94a5d9913ce72be8efb9f2333657e2c3292b04d748051bd0c"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.227143 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" event={"ID":"ada28834-1248-46cb-b482-4a2d8ee77520","Type":"ContainerStarted","Data":"174befe67a05a3b11727e199cd94f4cc6d1a445f04b6f132fc09c5ff5f7f136d"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.229246 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" event={"ID":"c5e2cc1d-7b01-4316-a960-5c47a420d6fd","Type":"ContainerStarted","Data":"c01c773d3a9b392414c75bcd72402eed0d22f14d3c83d2bfeefa7bd01a247370"} Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.229310 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" event={"ID":"c5e2cc1d-7b01-4316-a960-5c47a420d6fd","Type":"ContainerStarted","Data":"50cc43924fa76323f210a15e90fd72d1f0d968be448fc9383aa140aa4576b1aa"} Oct 09 10:19:30 crc kubenswrapper[4923]: E1009 10:19:30.233671 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" podUID="c5e2cc1d-7b01-4316-a960-5c47a420d6fd" Oct 09 10:19:30 crc kubenswrapper[4923]: I1009 10:19:30.236917 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" event={"ID":"6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed","Type":"ContainerStarted","Data":"fb74b72946951b7f9155fc1fcfcf55ee5c19994e9d972008c09050e8312ff6ea"} Oct 09 10:19:31 crc kubenswrapper[4923]: I1009 10:19:31.279147 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" event={"ID":"2ac22a8a-f293-4c68-85be-286bc263e77b","Type":"ContainerStarted","Data":"e6277f169358aaa88fc58fad3234ba608638ab9c980f919556ada1b5b8d3cf8a"} Oct 09 10:19:31 crc kubenswrapper[4923]: E1009 10:19:31.281536 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" podUID="c0d0a77b-c472-4690-b169-becf1ea318dc" Oct 09 10:19:31 crc kubenswrapper[4923]: E1009 10:19:31.281929 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" podUID="9e8c5a78-2a63-4219-8293-ea6d866a2d1b" Oct 09 10:19:31 crc kubenswrapper[4923]: E1009 10:19:31.282059 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" podUID="c5e2cc1d-7b01-4316-a960-5c47a420d6fd" Oct 09 10:19:31 crc kubenswrapper[4923]: E1009 10:19:31.282106 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:33652e75a03a058769019fe8d8c51585a6eeefef5e1ecb96f9965434117954f2\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" podUID="a44d21c5-6bf2-445e-9d72-2463571f1cf3" Oct 09 10:19:31 crc kubenswrapper[4923]: I1009 10:19:31.284615 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:31 crc kubenswrapper[4923]: E1009 10:19:31.303478 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" podUID="cd75b71e-8663-4e84-a99f-2d902135917d" Oct 09 10:19:31 crc kubenswrapper[4923]: I1009 10:19:31.319393 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" podStartSLOduration=5.319343219 podStartE2EDuration="5.319343219s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:19:31.316521921 +0000 UTC m=+857.384703677" watchObservedRunningTime="2025-10-09 10:19:31.319343219 +0000 UTC m=+857.387524975" Oct 09 10:19:38 crc kubenswrapper[4923]: I1009 10:19:38.065591 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-7d6957655c-h7b88" Oct 09 10:19:42 crc kubenswrapper[4923]: E1009 10:19:42.041447 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff" Oct 09 10:19:42 crc kubenswrapper[4923]: E1009 10:19:42.042648 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5d9g8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-68b6c87b68-2cvzn_openstack-operators(ada28834-1248-46cb-b482-4a2d8ee77520): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:19:42 crc kubenswrapper[4923]: E1009 10:19:42.601749 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351" Oct 09 10:19:42 crc kubenswrapper[4923]: E1009 10:19:42.602394 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent@sha256:03b4f3db4b373515f7e4095984b97197c05a14f87b2a0a525eb5d7be1d7bda66,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner@sha256:6722a752fb7cbffbae811f6ad6567120fbd4ebbe8c38a83ec2df02850a3276bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api@sha256:2115452234aedb505ed4efc6cd9b9a4ce3b9809aa7d0128d8fbeeee84dad1a69,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator@sha256:50597a8eaa6c4383f357574dcab8358b698729797b4156d932985a08ab86b7cd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener@sha256:cb4997d62c7b2534233a676cb92e19cf85dda07e2fb9fa642c28aab30489f69a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier@sha256:1ccbf3f6cf24c9ee91bed71467491e22b8cb4b95bce90250f4174fae936b0fa1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24@sha256:e91d58021b54c46883595ff66be65882de54abdb3be2ca53c4162b20d18b5f48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener@sha256:e7dcc3bf23d5e0393ac173e3c43d4ae85f4613a4fd16b3c147dc32ae491d49bf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker@sha256:2a1a8b582c6e4cc31081bd8b0887acf45e31c1d14596c4e361d27d08fef0debf,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:86daeb9c834bfcedb533086dff59a6b5b6e832b94ce2a9116337f8736bb80032,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute@sha256:5d4fdf424fad33a3650163e9e7423f92e97de3305508c2b7c6435822e0313189,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi@sha256:6d28de018f6e1672e775a75735e3bc16b63da41acd8fb5196ee0b06856c07133,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter@sha256:7211a617ec657701ca819aa0ba28e1d5750f5bf2c1391b755cc4a48cc360b0fa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification@sha256:c5fc9b72fc593bcf3b569c7ed24a256448eb1afab1504e668a3822e978be1306,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core@sha256:09b5017c95d7697e66b9c64846bc48ef5826a009cba89b956ec54561e5f4a2d1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup@sha256:88b99249f15470f359fb554f7f3a56974b743f4655e3f0c982c0260f75a67697,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler@sha256:e861d66785047d39eb68d9bac23e3f57ac84d9bd95593502d9b3b913b99fd1a4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume@sha256:b95f09bf3d259f9eacf3b63931977483f5c3c332f49b95ee8a69d8e3fb71d082,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api@sha256:6fc7801c0d18d41b9f11484b1cdb342de9cebd93072ec2205dbe40945715184f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9@sha256:d4d824b80cbed683543d9e8c7045ac97e080774f45a5067ccbca26404e067821,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central@sha256:182ec75938d8d3fb7d8f916373368add24062fec90489aa57776a81d0b36ea20,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns@sha256:9507ba5ab74cbae902e2dc07f89c7b3b5b76d8079e444365fe0eee6000fd7aaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer@sha256:17db080dcc4099f8a20aa0f238b6bca5c104672ae46743adeab9d1637725ecaa,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound@sha256:fd55cf3d73bfdc518419c9ba0b0cbef275140ae2d3bd0342a7310f81d57c2d78,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker@sha256:d164a9bd383f50df69fc22e7422f4650cd5076c90ed19278fc0f04e54345a63d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr@sha256:6beffe7d0bd75f9d1f495aeb7ab2334a2414af2c581d4833363df8441ed01018,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid@sha256:261e76f60c6bc6b172dc3608504552c63e83358a4fa3c0952a671544d83aa83f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler@sha256:581b65b646301e0fcb07582150ba63438f1353a85bf9acf1eb2acb4ce71c58bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron@sha256:2308c7b6c3d0aabbadfc9a06d84d67d2243f27fe8eed740ee96b1ce910203f62,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd@sha256:02d33f59749441cd5751c319e9d7cff97ab1004844c0e992650d340c6e8fbf43,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent@sha256:9cf0ca292340f1f978603955ef682effbf24316d6e2376b1c89906d84c3f06d0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn@sha256:c3e651f35b930bcf1a3084be8910c2f3f34d22a976c5379cf518a68d9994bfa7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent@sha256:58f678016d7f6c8fe579abe886fd138ef853642faa6766ca60639feac12d82ac,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent@sha256:46f92909153aaf03a585374b77d103c536509747e3270558d9a533295c46a7c5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter@sha256:39c642b2b337e38c18e80266fb14383754178202f40103646337722a594d984c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent@sha256:7fe367f51638c5c302fd3f8e66a31b09cb3b11519a7f72ef142b6c6fe8b91694,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter@sha256:d339ba049bbd1adccb795962bf163f5b22fd84dea865d88b9eb525e46247d6bd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:9ebf424d4107275a2e3f21f7a18ef257ff2f97c1298109ac7c802a5a4f4794f2,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api@sha256:4fcbe0d9a3c845708ecc32102ad4abbcbd947d87e5cf91f186de75b5d84ec681,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn@sha256:58a4e9a4dea86635c93ce37a2bb3c60ece62b3d656f6ee6a8845347cbb3e90fd,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine@sha256:6f2b843bc9f4ceb1ee873972d69e6bae6e1dbd378b486995bc3697d8bcff6339,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached@sha256:773daada6402d9cad089cdc809d6c0335456d057ac1a25441ab5d82add2f70f4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis@sha256:7323406a63fb3fdbb3eea4da0f7e8ed89c94c9bd0ad5ecd6c18fa4a4c2c550c4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api@sha256:7ae82068011e2d2e5ddc88c943fd32ff4a11902793e7a1df729811b2e27122a0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor@sha256:0c762c15d9d98d39cc9dc3d1f9a70f9188fef58d4e2f3b0c69c896cab8da5e48,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector@sha256:febf65561eeef5b36b70d0d65ee83f6451e43ec97bfab4d826e14215da6ff19b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent@sha256:b8aadfc3d547c5ef1e27fcb573d4760cf8c2f2271eefe1793c35a0d46b640837,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe@sha256:ecc91fd5079ee6d0c6ae1b11e97da790e33864d0e1930e574f959da2bddfa59a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent@sha256:2e981e93f99c929a3f04e5e41c8f645d44d390a9aeee3c5193cce7ec2edcbf3a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone@sha256:1e5714637b6e1a24c2858fe6d9bbb3f00bc61d69ad74a657b1c23682bf4cb2b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics@sha256:db384bf43222b066c378e77027a675d4cd9911107adba46c2922b3a55e10d6fb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api@sha256:35b8dcf27dc3b67f3840fa0e693ff312f74f7e22c634dff206a5c4d0133c716c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler@sha256:e109e4863e05e803dbfe04917756fd52231c560c65353170a2000be6cc2bb53d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share@sha256:6df0bebd9318ce11624413249e7e9781311638f276f8877668d3b382fe90e62f,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:56b75d97f4a48c8cf58b3a7c18c43618efb308bf0188124f6301142e61299b0c,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils@sha256:a51ed62767206067aa501142dbf01f20b3d65325d30faf1b4d6424d5b17dfba5,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api@sha256:592e3cd32d3cc97a69093ad905b449aa374ffbb1b2644b738bb6c1434476d1f6,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute@sha256:5f179b847f2dc32d9110b8f2be9fe65f1aeada1e18105dffdaf052981215d844,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor@sha256:9596452e283febbe08204d0ef0fd1992af3395d0969f7ac76663ed7c8be5b4d4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy@sha256:d61005a10bef1b37762a8a41e6755c1169241e36cc5f92886bca6f4f6b9c381a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler@sha256:e6a4335bcbeed3cd3e73ac879f754e314761e4a417a67539ca88e96a79346328,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api@sha256:97d88fc53421b699fc91983313d7beec4a0f177089e95bdf5ba15c3f521db9a9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager@sha256:5365e5c9c3ad2ede1b6945255b2cc6b009d642c39babdf25e0655282cfa646fe,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping@sha256:5b55795d774e0ea160ff8a7fd491ed41cf2d93c7d821694abb3a879eaffcefeb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog@sha256:26e955c46a6063eafcfeb79430bf3d9268dbe95687c00e63a624b3ec5a846f5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker@sha256:58939baa18ab09e2b24996c5f3665ae52274b781f661ea06a67c991e9a832d5a,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient@sha256:b8bff6857fec93c3c1521f1a8c23de21bcb86fc0f960972e81f6c3f95d4185be,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather@sha256:943eee724277e252795909137538a553ef5284c8103ad01b9be7b0138c66d14d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter@sha256:ecd56e6733c475f2d441344fd98f288c3eac0261ba113695fec7520a954ccbc7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi@sha256:d97b08fd421065c8c33a523973822ac468500cbe853069aa9214393fbda7a908,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller@sha256:d76f7d6620930cc2e9ac070492bbeb525f83ce5ff4947463e3784bf1ce04a857,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base@sha256:289dea3beea1cd4405895fc42e44372b35e4a941e31c59e102c333471a3ca9b7,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server@sha256:9b19894fa67a81bf8ba4159b55b49f38877c670aeb97e2021c341cef2a9294e4,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd@sha256:ea164961ad30453ad0301c6b73364e1f1024f689634c88dd98265f9c7048e31d,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server@sha256:6f9f2ea45f0271f6da8eb05a5f74cf5ce6769479346f5c2f407ee6f31a9c7ff3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:59448516174fc3bab679b9a8dd62cb9a9d16b5734aadbeb98e960e3b7c79bd22,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:adcdeb8ecd601fb03c3b0901d5b5111af2ca48f7dd443e22224db6daaf08f5d0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account@sha256:2bf32d9b95899d7637dfe19d07cf1ecc9a06593984faff57a3c0dce060012edb,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container@sha256:7a452cd18b64d522e8a1e25bdcea543e9fe5f5b76e1c5e044c2b5334e06a326b,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object@sha256:6a46aa13aa359b8e782a22d67db42db02bbf2bb7e35df4b684ac1daeda38cde3,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server@sha256:f6824854bea6b2acbb00c34639799b4744818d4adbdd40e37dc5088f9ae18d58,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all@sha256:a66d2fdc21f25c690f02e643d2666dbe7df43a64cd55086ec33d6755e6d809b9,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api@sha256:e0e84e3fb8f9f12e9280d3d28b415639abfad5b5e46787a61b42beb97f32c76e,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier@sha256:bd5376b807eca8a409ea663906d9a990cd95a109b8f1f2c56c06f9201be1f6ca,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine@sha256:366de9fadc79d3ea77de93ef1c81e4b8525764f0704e3f30d1a8afa2745c8166,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-98sfh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-677c5f5bffmcv74_openstack-operators(29e93b24-6e22-4ade-9fd9-1742186d21fd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:19:42 crc kubenswrapper[4923]: E1009 10:19:42.898579 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" podUID="29e93b24-6e22-4ade-9fd9-1742186d21fd" Oct 09 10:19:43 crc kubenswrapper[4923]: E1009 10:19:43.284377 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" podUID="ada28834-1248-46cb-b482-4a2d8ee77520" Oct 09 10:19:43 crc kubenswrapper[4923]: I1009 10:19:43.397375 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" event={"ID":"4466effc-4c4f-4023-94bc-d1e2f48d5d3b","Type":"ContainerStarted","Data":"20576567ed5cc4921cbeac01f3ddfcf603b5cb51c5a3fb086970042f908d4b67"} Oct 09 10:19:43 crc kubenswrapper[4923]: I1009 10:19:43.398701 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" event={"ID":"42a720da-4703-4d8b-9046-6c8e1643c880","Type":"ContainerStarted","Data":"2b3ce4786912fe1e103b39f07143b93bdce65da614b1a4cfb09981ce39acb170"} Oct 09 10:19:43 crc kubenswrapper[4923]: I1009 10:19:43.400519 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" event={"ID":"29e93b24-6e22-4ade-9fd9-1742186d21fd","Type":"ContainerStarted","Data":"efd2bfb70feb9a655b1900753b4c77afd5597043fc77163110942c39965cf4ba"} Oct 09 10:19:43 crc kubenswrapper[4923]: I1009 10:19:43.407139 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" event={"ID":"ada28834-1248-46cb-b482-4a2d8ee77520","Type":"ContainerStarted","Data":"97ab1c4a92b2b732e1aa76a5111ecc41ecb0f35aac70a9775c15c33788555d15"} Oct 09 10:19:43 crc kubenswrapper[4923]: E1009 10:19:43.410028 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" podUID="29e93b24-6e22-4ade-9fd9-1742186d21fd" Oct 09 10:19:43 crc kubenswrapper[4923]: E1009 10:19:43.410033 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" podUID="ada28834-1248-46cb-b482-4a2d8ee77520" Oct 09 10:19:43 crc kubenswrapper[4923]: I1009 10:19:43.415917 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" event={"ID":"91f871ca-9617-49c0-877f-ee0bcf938bee","Type":"ContainerStarted","Data":"15149a59b60b76ba880c6eee5bc12174dee97314ad946c0d62aef0d851d41f44"} Oct 09 10:19:43 crc kubenswrapper[4923]: I1009 10:19:43.417585 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" event={"ID":"3fd48b51-1de2-4d9f-a790-55c95638fc2f","Type":"ContainerStarted","Data":"d2f05b39e818b68cb1be205485aa7a15cedcfa62d95e33e80d589eacd2240c25"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.437499 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" event={"ID":"36c97947-5f3f-4d0e-8125-0663f04f9eaf","Type":"ContainerStarted","Data":"a7e32f3d1f76a1007df80f6f29b47fd06184dd8f386dea12aabd11bf8c664092"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.440469 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" event={"ID":"120ce970-aff3-40cb-8724-957a034fb914","Type":"ContainerStarted","Data":"32b31654c9f326057bb80584eba254750640ad6e24e77879a1c9cb25ce5ca008"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.441842 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" event={"ID":"575ef7c8-b9ea-47ae-aa33-8c87596fff08","Type":"ContainerStarted","Data":"711bf77943f99f7672ba16819081f5962c97d5b48463f7d2a186052d61930efa"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.446086 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" event={"ID":"4466effc-4c4f-4023-94bc-d1e2f48d5d3b","Type":"ContainerStarted","Data":"3ddf6f072b98c7cce44b12d02bfc3f64bf561856b880a89d8a34fb85d7520eb4"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.446163 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.448400 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" event={"ID":"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8","Type":"ContainerStarted","Data":"32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.450807 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" event={"ID":"4609ba20-418f-4134-b671-10154112f387","Type":"ContainerStarted","Data":"4b8a58b2bbd5abc74949982d9dd6c2772f6e3ca3e006acbc596d957c3c2f9355"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.455249 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" event={"ID":"1438c7f5-06a5-49d0-be96-7fbee10b3010","Type":"ContainerStarted","Data":"006e044f5e23d5e492d9cdf33531f19c158f53ab72e18422355545980f4d50f4"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.455286 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" event={"ID":"1438c7f5-06a5-49d0-be96-7fbee10b3010","Type":"ContainerStarted","Data":"b3f284028fdf5cfe226bd653b69de615ac4f6514cbffe3423e0ef87c10671358"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.455409 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.461421 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" event={"ID":"c5fba54a-5e97-419a-8950-65bd3908aadb","Type":"ContainerStarted","Data":"1efd4f79770a1d1cc375b223714593e20abfe01a56671d2162116620521cb725"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.472023 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" event={"ID":"3fd48b51-1de2-4d9f-a790-55c95638fc2f","Type":"ContainerStarted","Data":"8176ccb20ae07cd354dc4a9998786fa3b22a0fd8b2f19d2b379ad3c1742703fb"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.472978 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.478385 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" podStartSLOduration=3.941390207 podStartE2EDuration="18.478363105s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.179244993 +0000 UTC m=+854.247426749" lastFinishedPulling="2025-10-09 10:19:42.716217891 +0000 UTC m=+868.784399647" observedRunningTime="2025-10-09 10:19:44.472329296 +0000 UTC m=+870.540511052" watchObservedRunningTime="2025-10-09 10:19:44.478363105 +0000 UTC m=+870.546544861" Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.490805 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" event={"ID":"42a720da-4703-4d8b-9046-6c8e1643c880","Type":"ContainerStarted","Data":"0e7092d424b0b5bc0744893660f8825cf35078836176f6d97f4e2cda75bb6267"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.491173 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.508875 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" podStartSLOduration=3.9169879979999997 podStartE2EDuration="18.508858032s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.125715605 +0000 UTC m=+854.193897361" lastFinishedPulling="2025-10-09 10:19:42.717585639 +0000 UTC m=+868.785767395" observedRunningTime="2025-10-09 10:19:44.508119742 +0000 UTC m=+870.576301498" watchObservedRunningTime="2025-10-09 10:19:44.508858032 +0000 UTC m=+870.577039788" Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.510191 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" event={"ID":"a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8","Type":"ContainerStarted","Data":"44fd26fab70596f3fe9dc756575b62c83e65e6024d82fe30e40d39a471512bdf"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.535412 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" event={"ID":"c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72","Type":"ContainerStarted","Data":"aa207434fdf8a1e26462e8f9c5e79ff23dd9a96c71b0f28c155d1fca6e9d55fe"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.555281 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" podStartSLOduration=4.561649797 podStartE2EDuration="18.555253652s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.723300036 +0000 UTC m=+854.791481792" lastFinishedPulling="2025-10-09 10:19:42.716903891 +0000 UTC m=+868.785085647" observedRunningTime="2025-10-09 10:19:44.550973602 +0000 UTC m=+870.619155358" watchObservedRunningTime="2025-10-09 10:19:44.555253652 +0000 UTC m=+870.623435408" Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.560813 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" event={"ID":"6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed","Type":"ContainerStarted","Data":"348203c6b51dc7552814c23923bf809669c9b0d905213f3c3b7610497fce68dc"} Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.565220 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" event={"ID":"2ad61b43-c2ba-4dc9-8376-e75817b9e153","Type":"ContainerStarted","Data":"effe70131ad51ed0ada1eade5bc54b4d2d3e2453057fe29ed818def7ecc333c8"} Oct 09 10:19:44 crc kubenswrapper[4923]: E1009 10:19:44.567479 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" podUID="29e93b24-6e22-4ade-9fd9-1742186d21fd" Oct 09 10:19:44 crc kubenswrapper[4923]: E1009 10:19:44.567845 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" podUID="ada28834-1248-46cb-b482-4a2d8ee77520" Oct 09 10:19:44 crc kubenswrapper[4923]: I1009 10:19:44.641862 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" podStartSLOduration=4.653916934 podStartE2EDuration="18.641831309s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.727258197 +0000 UTC m=+854.795439953" lastFinishedPulling="2025-10-09 10:19:42.715172562 +0000 UTC m=+868.783354328" observedRunningTime="2025-10-09 10:19:44.578687563 +0000 UTC m=+870.646869319" watchObservedRunningTime="2025-10-09 10:19:44.641831309 +0000 UTC m=+870.710013065" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.573630 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" event={"ID":"36c97947-5f3f-4d0e-8125-0663f04f9eaf","Type":"ContainerStarted","Data":"c6206d8b8069cdab382943a8ff7a7f790eac1fa3f5f29eb1461aa96eb78fc4a9"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.575356 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.580666 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" event={"ID":"575ef7c8-b9ea-47ae-aa33-8c87596fff08","Type":"ContainerStarted","Data":"681a0d36fc899d550b38cdde8b9961e6b2b0180ea43cfd74ac6f67083a09c77b"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.581293 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.592788 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" event={"ID":"c5fba54a-5e97-419a-8950-65bd3908aadb","Type":"ContainerStarted","Data":"e2fe04784e6fa758f8b0e8d3cbdec88de7da26250bb0071944307a98eb9ecdef"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.592926 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.598074 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" event={"ID":"a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8","Type":"ContainerStarted","Data":"81a465d5a50ce73c32db76c4cfeac8aabef3eecd7086e255b185db8913ef20c0"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.598138 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.600204 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" event={"ID":"c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72","Type":"ContainerStarted","Data":"3dba5ca6151a2b683cf21fab62bf4630efe9e00ddc09e9cdceb1e48adf25f110"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.600407 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.613037 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" event={"ID":"4609ba20-418f-4134-b671-10154112f387","Type":"ContainerStarted","Data":"25c64ad59a82b0dabe2a266b8bd2436f07988bb8a720c58aff3a11f3e48f3062"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.613212 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.614531 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" podStartSLOduration=5.584680066 podStartE2EDuration="19.614505646s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.684292133 +0000 UTC m=+854.752473889" lastFinishedPulling="2025-10-09 10:19:42.714117703 +0000 UTC m=+868.782299469" observedRunningTime="2025-10-09 10:19:45.604297472 +0000 UTC m=+871.672479228" watchObservedRunningTime="2025-10-09 10:19:45.614505646 +0000 UTC m=+871.682687402" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.618310 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" event={"ID":"2ad61b43-c2ba-4dc9-8376-e75817b9e153","Type":"ContainerStarted","Data":"810cc5980182de5e537e61b3b822ed64a4b99a37d6cbc0fe05448e25b3a873e3"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.618960 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.630940 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" event={"ID":"6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed","Type":"ContainerStarted","Data":"8845ffe285259bad91ec7ccabd8629774932a7c807de16b416f8d0eb31836624"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.631206 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.643461 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" event={"ID":"120ce970-aff3-40cb-8724-957a034fb914","Type":"ContainerStarted","Data":"8cace1cda8dfc1d9293f67a853a5485860c476203ca9535152900fd3f9be0c4b"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.644338 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.646868 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" event={"ID":"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8","Type":"ContainerStarted","Data":"e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.647027 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.649899 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" event={"ID":"91f871ca-9617-49c0-877f-ee0bcf938bee","Type":"ContainerStarted","Data":"7724ff143a079fb759dec844d802e1be2090a9396afbf0b871ba9b736941299b"} Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.650119 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.656774 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" podStartSLOduration=5.630329064 podStartE2EDuration="19.656726859s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.692766739 +0000 UTC m=+854.760948495" lastFinishedPulling="2025-10-09 10:19:42.719164524 +0000 UTC m=+868.787346290" observedRunningTime="2025-10-09 10:19:45.651298049 +0000 UTC m=+871.719479815" watchObservedRunningTime="2025-10-09 10:19:45.656726859 +0000 UTC m=+871.724908615" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.660946 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" podStartSLOduration=5.671560051 podStartE2EDuration="19.660924877s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.728573183 +0000 UTC m=+854.796754939" lastFinishedPulling="2025-10-09 10:19:42.717937979 +0000 UTC m=+868.786119765" observedRunningTime="2025-10-09 10:19:45.630926343 +0000 UTC m=+871.699108099" watchObservedRunningTime="2025-10-09 10:19:45.660924877 +0000 UTC m=+871.729106633" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.682173 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" podStartSLOduration=5.554087965 podStartE2EDuration="19.682149896s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.590492985 +0000 UTC m=+854.658674741" lastFinishedPulling="2025-10-09 10:19:42.718554916 +0000 UTC m=+868.786736672" observedRunningTime="2025-10-09 10:19:45.673867906 +0000 UTC m=+871.742049672" watchObservedRunningTime="2025-10-09 10:19:45.682149896 +0000 UTC m=+871.750331652" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.727174 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" podStartSLOduration=5.731395953 podStartE2EDuration="19.727156527s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.721545488 +0000 UTC m=+854.789727244" lastFinishedPulling="2025-10-09 10:19:42.717306052 +0000 UTC m=+868.785487818" observedRunningTime="2025-10-09 10:19:45.704504848 +0000 UTC m=+871.772686604" watchObservedRunningTime="2025-10-09 10:19:45.727156527 +0000 UTC m=+871.795338283" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.727497 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" podStartSLOduration=5.158528629 podStartE2EDuration="19.727492356s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.146917135 +0000 UTC m=+854.215098891" lastFinishedPulling="2025-10-09 10:19:42.715880852 +0000 UTC m=+868.784062618" observedRunningTime="2025-10-09 10:19:45.723539987 +0000 UTC m=+871.791721753" watchObservedRunningTime="2025-10-09 10:19:45.727492356 +0000 UTC m=+871.795674102" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.784148 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" podStartSLOduration=6.34069156 podStartE2EDuration="19.784127261s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.289986099 +0000 UTC m=+855.358167855" lastFinishedPulling="2025-10-09 10:19:42.73342178 +0000 UTC m=+868.801603556" observedRunningTime="2025-10-09 10:19:45.779328368 +0000 UTC m=+871.847510134" watchObservedRunningTime="2025-10-09 10:19:45.784127261 +0000 UTC m=+871.852309017" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.785035 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" podStartSLOduration=6.409540915 podStartE2EDuration="19.785027136s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.339745173 +0000 UTC m=+855.407926929" lastFinishedPulling="2025-10-09 10:19:42.715231394 +0000 UTC m=+868.783413150" observedRunningTime="2025-10-09 10:19:45.756502703 +0000 UTC m=+871.824684459" watchObservedRunningTime="2025-10-09 10:19:45.785027136 +0000 UTC m=+871.853208892" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.818468 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" podStartSLOduration=6.250307459 podStartE2EDuration="19.818448745s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.147873869 +0000 UTC m=+855.216055625" lastFinishedPulling="2025-10-09 10:19:42.716015135 +0000 UTC m=+868.784196911" observedRunningTime="2025-10-09 10:19:45.816083599 +0000 UTC m=+871.884265365" watchObservedRunningTime="2025-10-09 10:19:45.818448745 +0000 UTC m=+871.886630501" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.844998 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" podStartSLOduration=4.8142151779999995 podStartE2EDuration="19.844973582s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:27.685513459 +0000 UTC m=+853.753695215" lastFinishedPulling="2025-10-09 10:19:42.716271863 +0000 UTC m=+868.784453619" observedRunningTime="2025-10-09 10:19:45.838581095 +0000 UTC m=+871.906762861" watchObservedRunningTime="2025-10-09 10:19:45.844973582 +0000 UTC m=+871.913155338" Oct 09 10:19:45 crc kubenswrapper[4923]: I1009 10:19:45.859990 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" podStartSLOduration=5.796163034 podStartE2EDuration="19.85996974s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:28.656009236 +0000 UTC m=+854.724190992" lastFinishedPulling="2025-10-09 10:19:42.719815932 +0000 UTC m=+868.787997698" observedRunningTime="2025-10-09 10:19:45.854155278 +0000 UTC m=+871.922337044" watchObservedRunningTime="2025-10-09 10:19:45.85996974 +0000 UTC m=+871.928151496" Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.696072 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" event={"ID":"c0d0a77b-c472-4690-b169-becf1ea318dc","Type":"ContainerStarted","Data":"3085ba200b18251337b6b0e22b7795df5117e8ef4e6e30abf70538341068ae66"} Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.698025 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.702010 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" event={"ID":"a44d21c5-6bf2-445e-9d72-2463571f1cf3","Type":"ContainerStarted","Data":"50793930fd923fa26a4acbe181e09b9930f8f22652a93e83cc1cbbda8b998295"} Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.702692 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.705664 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" event={"ID":"9e8c5a78-2a63-4219-8293-ea6d866a2d1b","Type":"ContainerStarted","Data":"782e612923465af760808f5c41108975cbe4a117bcec8274cc82abb46ef869db"} Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.705966 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.708208 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" event={"ID":"c5e2cc1d-7b01-4316-a960-5c47a420d6fd","Type":"ContainerStarted","Data":"909201427b213b480ac043decc38e50763411de68ad8f4c74498de83c6804019"} Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.708499 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.710271 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" event={"ID":"cd75b71e-8663-4e84-a99f-2d902135917d","Type":"ContainerStarted","Data":"890d9f9fcaaae0e5b6f5f23d0562914171401ffc8a0787ab967d1460130a3c32"} Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.718149 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" podStartSLOduration=4.26202637 podStartE2EDuration="23.718133856s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.313145573 +0000 UTC m=+855.381327329" lastFinishedPulling="2025-10-09 10:19:48.769253059 +0000 UTC m=+874.837434815" observedRunningTime="2025-10-09 10:19:49.718119865 +0000 UTC m=+875.786301661" watchObservedRunningTime="2025-10-09 10:19:49.718133856 +0000 UTC m=+875.786315612" Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.739690 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t" podStartSLOduration=4.331623475 podStartE2EDuration="23.739669405s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.379616311 +0000 UTC m=+855.447798067" lastFinishedPulling="2025-10-09 10:19:48.787662241 +0000 UTC m=+874.855843997" observedRunningTime="2025-10-09 10:19:49.733303087 +0000 UTC m=+875.801484853" watchObservedRunningTime="2025-10-09 10:19:49.739669405 +0000 UTC m=+875.807851181" Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.754373 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" podStartSLOduration=4.332363324 podStartE2EDuration="23.754352842s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.379853117 +0000 UTC m=+855.448034873" lastFinishedPulling="2025-10-09 10:19:48.801842645 +0000 UTC m=+874.870024391" observedRunningTime="2025-10-09 10:19:49.7488556 +0000 UTC m=+875.817037376" watchObservedRunningTime="2025-10-09 10:19:49.754352842 +0000 UTC m=+875.822534618" Oct 09 10:19:49 crc kubenswrapper[4923]: I1009 10:19:49.772512 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" podStartSLOduration=4.350168149 podStartE2EDuration="23.772495556s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.344594477 +0000 UTC m=+855.412776233" lastFinishedPulling="2025-10-09 10:19:48.766921884 +0000 UTC m=+874.835103640" observedRunningTime="2025-10-09 10:19:49.769498793 +0000 UTC m=+875.837680549" watchObservedRunningTime="2025-10-09 10:19:49.772495556 +0000 UTC m=+875.840677312" Oct 09 10:19:56 crc kubenswrapper[4923]: I1009 10:19:56.405555 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-658bdf4b74-h294t" Oct 09 10:19:56 crc kubenswrapper[4923]: I1009 10:19:56.433075 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" podStartSLOduration=11.02317481 podStartE2EDuration="30.433056322s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.313494062 +0000 UTC m=+855.381675818" lastFinishedPulling="2025-10-09 10:19:48.723375574 +0000 UTC m=+874.791557330" observedRunningTime="2025-10-09 10:19:49.801862033 +0000 UTC m=+875.870043799" watchObservedRunningTime="2025-10-09 10:19:56.433056322 +0000 UTC m=+882.501238078" Oct 09 10:19:56 crc kubenswrapper[4923]: I1009 10:19:56.498408 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-7b7fb68549-lpvgj" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:56.508990 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-85d5d9dd78-lcs47" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:56.555625 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-84b9b84486-fhsk4" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:56.562430 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-858f76bbdd-zjhgq" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:56.585385 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-7ffbcb7588-gdprs" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:56.865526 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5f67fbc655-jz6s9" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:56.905545 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-f9fb45f8f-md5f2" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:56.956648 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-9c5c78d49-n8tdz" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:56.966127 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-55b6b7c7b8-qrh4j" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:57.024549 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-79d585cb66-tp6xl" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:57.024598 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-69fdcfc5f5-pbjk9" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:57.057459 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5df598886f-55fjr" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:57.198101 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-656bcbd775-td8lq" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:57.437701 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-79db49b9fb-f89mg" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:57.633767 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-db6d7f97b-xzpv6" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:57.700336 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-67cfc6749b-hw8r6" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:57.704390 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" Oct 09 10:20:03 crc kubenswrapper[4923]: I1009 10:19:57.732744 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-7f554bff7b-54cbp" Oct 09 10:20:09 crc kubenswrapper[4923]: I1009 10:20:09.859352 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" event={"ID":"ada28834-1248-46cb-b482-4a2d8ee77520","Type":"ContainerStarted","Data":"5239bb1a655d5637d96d0dfed892d357063635f2cda0e858880f769de7c66d6b"} Oct 09 10:20:09 crc kubenswrapper[4923]: I1009 10:20:09.860304 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" Oct 09 10:20:09 crc kubenswrapper[4923]: I1009 10:20:09.863201 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" event={"ID":"29e93b24-6e22-4ade-9fd9-1742186d21fd","Type":"ContainerStarted","Data":"cb093d45389dd91d784d62c6d140d647154348efd476c43bbb9f2459e19e29b4"} Oct 09 10:20:09 crc kubenswrapper[4923]: I1009 10:20:09.863425 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:20:09 crc kubenswrapper[4923]: I1009 10:20:09.895129 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" podStartSLOduration=3.629706723 podStartE2EDuration="43.89510678s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.136683188 +0000 UTC m=+855.204864944" lastFinishedPulling="2025-10-09 10:20:09.402083245 +0000 UTC m=+895.470265001" observedRunningTime="2025-10-09 10:20:09.883625711 +0000 UTC m=+895.951807467" watchObservedRunningTime="2025-10-09 10:20:09.89510678 +0000 UTC m=+895.963288536" Oct 09 10:20:09 crc kubenswrapper[4923]: I1009 10:20:09.924663 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" podStartSLOduration=3.661509716 podStartE2EDuration="43.924634281s" podCreationTimestamp="2025-10-09 10:19:26 +0000 UTC" firstStartedPulling="2025-10-09 10:19:29.136664227 +0000 UTC m=+855.204845983" lastFinishedPulling="2025-10-09 10:20:09.399788802 +0000 UTC m=+895.467970548" observedRunningTime="2025-10-09 10:20:09.922724658 +0000 UTC m=+895.990906414" watchObservedRunningTime="2025-10-09 10:20:09.924634281 +0000 UTC m=+895.992816037" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.347399 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wpz6r"] Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.351344 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.374819 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wpz6r"] Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.461995 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-utilities\") pod \"certified-operators-wpz6r\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.462090 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwl7r\" (UniqueName: \"kubernetes.io/projected/480ce995-be09-4597-8c1d-ea56c5ebe635-kube-api-access-gwl7r\") pod \"certified-operators-wpz6r\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.462160 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-catalog-content\") pod \"certified-operators-wpz6r\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.563797 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-utilities\") pod \"certified-operators-wpz6r\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.563868 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwl7r\" (UniqueName: \"kubernetes.io/projected/480ce995-be09-4597-8c1d-ea56c5ebe635-kube-api-access-gwl7r\") pod \"certified-operators-wpz6r\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.563934 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-catalog-content\") pod \"certified-operators-wpz6r\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.564451 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-catalog-content\") pod \"certified-operators-wpz6r\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.564470 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-utilities\") pod \"certified-operators-wpz6r\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.585227 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwl7r\" (UniqueName: \"kubernetes.io/projected/480ce995-be09-4597-8c1d-ea56c5ebe635-kube-api-access-gwl7r\") pod \"certified-operators-wpz6r\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.678614 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:10 crc kubenswrapper[4923]: I1009 10:20:10.930097 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wpz6r"] Oct 09 10:20:11 crc kubenswrapper[4923]: I1009 10:20:11.883641 4923 generic.go:334] "Generic (PLEG): container finished" podID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerID="5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf" exitCode=0 Oct 09 10:20:11 crc kubenswrapper[4923]: I1009 10:20:11.883701 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpz6r" event={"ID":"480ce995-be09-4597-8c1d-ea56c5ebe635","Type":"ContainerDied","Data":"5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf"} Oct 09 10:20:11 crc kubenswrapper[4923]: I1009 10:20:11.883767 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpz6r" event={"ID":"480ce995-be09-4597-8c1d-ea56c5ebe635","Type":"ContainerStarted","Data":"5b95948b85be0421574d63ff6fa24433d4cc3a9c2fe7f47a30e1d4ce26f81c35"} Oct 09 10:20:14 crc kubenswrapper[4923]: I1009 10:20:14.915987 4923 generic.go:334] "Generic (PLEG): container finished" podID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerID="4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a" exitCode=0 Oct 09 10:20:14 crc kubenswrapper[4923]: I1009 10:20:14.916208 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpz6r" event={"ID":"480ce995-be09-4597-8c1d-ea56c5ebe635","Type":"ContainerDied","Data":"4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a"} Oct 09 10:20:15 crc kubenswrapper[4923]: I1009 10:20:15.928345 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpz6r" event={"ID":"480ce995-be09-4597-8c1d-ea56c5ebe635","Type":"ContainerStarted","Data":"6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc"} Oct 09 10:20:15 crc kubenswrapper[4923]: I1009 10:20:15.957902 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wpz6r" podStartSLOduration=2.459960047 podStartE2EDuration="5.957873459s" podCreationTimestamp="2025-10-09 10:20:10 +0000 UTC" firstStartedPulling="2025-10-09 10:20:11.886702511 +0000 UTC m=+897.954884307" lastFinishedPulling="2025-10-09 10:20:15.384615953 +0000 UTC m=+901.452797719" observedRunningTime="2025-10-09 10:20:15.948698753 +0000 UTC m=+902.016880519" watchObservedRunningTime="2025-10-09 10:20:15.957873459 +0000 UTC m=+902.026055225" Oct 09 10:20:17 crc kubenswrapper[4923]: I1009 10:20:17.532644 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-68b6c87b68-2cvzn" Oct 09 10:20:17 crc kubenswrapper[4923]: I1009 10:20:17.776059 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-677c5f5bffmcv74" Oct 09 10:20:20 crc kubenswrapper[4923]: I1009 10:20:20.679023 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:20 crc kubenswrapper[4923]: I1009 10:20:20.679681 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:20 crc kubenswrapper[4923]: I1009 10:20:20.741509 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:21 crc kubenswrapper[4923]: I1009 10:20:21.029789 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:21 crc kubenswrapper[4923]: I1009 10:20:21.074973 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wpz6r"] Oct 09 10:20:22 crc kubenswrapper[4923]: I1009 10:20:22.991212 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wpz6r" podUID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerName="registry-server" containerID="cri-o://6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc" gracePeriod=2 Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.425527 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.584288 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-catalog-content\") pod \"480ce995-be09-4597-8c1d-ea56c5ebe635\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.584535 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwl7r\" (UniqueName: \"kubernetes.io/projected/480ce995-be09-4597-8c1d-ea56c5ebe635-kube-api-access-gwl7r\") pod \"480ce995-be09-4597-8c1d-ea56c5ebe635\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.584591 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-utilities\") pod \"480ce995-be09-4597-8c1d-ea56c5ebe635\" (UID: \"480ce995-be09-4597-8c1d-ea56c5ebe635\") " Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.585390 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-utilities" (OuterVolumeSpecName: "utilities") pod "480ce995-be09-4597-8c1d-ea56c5ebe635" (UID: "480ce995-be09-4597-8c1d-ea56c5ebe635"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.595169 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480ce995-be09-4597-8c1d-ea56c5ebe635-kube-api-access-gwl7r" (OuterVolumeSpecName: "kube-api-access-gwl7r") pod "480ce995-be09-4597-8c1d-ea56c5ebe635" (UID: "480ce995-be09-4597-8c1d-ea56c5ebe635"). InnerVolumeSpecName "kube-api-access-gwl7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.644820 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "480ce995-be09-4597-8c1d-ea56c5ebe635" (UID: "480ce995-be09-4597-8c1d-ea56c5ebe635"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.686467 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwl7r\" (UniqueName: \"kubernetes.io/projected/480ce995-be09-4597-8c1d-ea56c5ebe635-kube-api-access-gwl7r\") on node \"crc\" DevicePath \"\"" Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.686530 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:20:23 crc kubenswrapper[4923]: I1009 10:20:23.686548 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/480ce995-be09-4597-8c1d-ea56c5ebe635-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.004689 4923 generic.go:334] "Generic (PLEG): container finished" podID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerID="6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc" exitCode=0 Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.004820 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpz6r" event={"ID":"480ce995-be09-4597-8c1d-ea56c5ebe635","Type":"ContainerDied","Data":"6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc"} Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.004868 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpz6r" event={"ID":"480ce995-be09-4597-8c1d-ea56c5ebe635","Type":"ContainerDied","Data":"5b95948b85be0421574d63ff6fa24433d4cc3a9c2fe7f47a30e1d4ce26f81c35"} Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.004907 4923 scope.go:117] "RemoveContainer" containerID="6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.004947 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wpz6r" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.031559 4923 scope.go:117] "RemoveContainer" containerID="4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.056620 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wpz6r"] Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.063126 4923 scope.go:117] "RemoveContainer" containerID="5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.064614 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wpz6r"] Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.092774 4923 scope.go:117] "RemoveContainer" containerID="6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc" Oct 09 10:20:24 crc kubenswrapper[4923]: E1009 10:20:24.093501 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc\": container with ID starting with 6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc not found: ID does not exist" containerID="6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.093548 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc"} err="failed to get container status \"6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc\": rpc error: code = NotFound desc = could not find container \"6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc\": container with ID starting with 6cefd91847a89ab15de1335f8f41b58135c6e6c4fcbb8c348b7325507482e8dc not found: ID does not exist" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.093577 4923 scope.go:117] "RemoveContainer" containerID="4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a" Oct 09 10:20:24 crc kubenswrapper[4923]: E1009 10:20:24.094473 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a\": container with ID starting with 4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a not found: ID does not exist" containerID="4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.094573 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a"} err="failed to get container status \"4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a\": rpc error: code = NotFound desc = could not find container \"4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a\": container with ID starting with 4cb9a080efc25bf0666b80494ab36f7c7b2dad74fa2606261162c643865c6c6a not found: ID does not exist" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.094648 4923 scope.go:117] "RemoveContainer" containerID="5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf" Oct 09 10:20:24 crc kubenswrapper[4923]: E1009 10:20:24.095482 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf\": container with ID starting with 5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf not found: ID does not exist" containerID="5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.095555 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf"} err="failed to get container status \"5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf\": rpc error: code = NotFound desc = could not find container \"5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf\": container with ID starting with 5dc137c61a782f935f4700d4f34474c5dcf65227a9f9ab38471318753f778cbf not found: ID does not exist" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.599818 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.599884 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:20:24 crc kubenswrapper[4923]: I1009 10:20:24.610527 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="480ce995-be09-4597-8c1d-ea56c5ebe635" path="/var/lib/kubelet/pods/480ce995-be09-4597-8c1d-ea56c5ebe635/volumes" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.971168 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-x9mlp"] Oct 09 10:20:33 crc kubenswrapper[4923]: E1009 10:20:33.972239 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerName="extract-utilities" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.972257 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerName="extract-utilities" Oct 09 10:20:33 crc kubenswrapper[4923]: E1009 10:20:33.972291 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerName="extract-content" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.972299 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerName="extract-content" Oct 09 10:20:33 crc kubenswrapper[4923]: E1009 10:20:33.972314 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerName="registry-server" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.972324 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerName="registry-server" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.972533 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="480ce995-be09-4597-8c1d-ea56c5ebe635" containerName="registry-server" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.973493 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.976831 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.976831 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.976952 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-f6gxs" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.977645 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 09 10:20:33 crc kubenswrapper[4923]: I1009 10:20:33.992262 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-x9mlp"] Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.046082 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ccck9"] Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.051239 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.055914 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.062121 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ccck9"] Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.154540 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfqlz\" (UniqueName: \"kubernetes.io/projected/bfe01f2f-670f-4c70-8139-f2988dd1bda9-kube-api-access-wfqlz\") pod \"dnsmasq-dns-7bfcb9d745-x9mlp\" (UID: \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.154603 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfe01f2f-670f-4c70-8139-f2988dd1bda9-config\") pod \"dnsmasq-dns-7bfcb9d745-x9mlp\" (UID: \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.154634 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6tg4\" (UniqueName: \"kubernetes.io/projected/ec75d180-568e-42d8-bb4c-4f82d3fa444b-kube-api-access-b6tg4\") pod \"dnsmasq-dns-758b79db4c-ccck9\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.154657 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-config\") pod \"dnsmasq-dns-758b79db4c-ccck9\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.154679 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-dns-svc\") pod \"dnsmasq-dns-758b79db4c-ccck9\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.255924 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfqlz\" (UniqueName: \"kubernetes.io/projected/bfe01f2f-670f-4c70-8139-f2988dd1bda9-kube-api-access-wfqlz\") pod \"dnsmasq-dns-7bfcb9d745-x9mlp\" (UID: \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.256322 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfe01f2f-670f-4c70-8139-f2988dd1bda9-config\") pod \"dnsmasq-dns-7bfcb9d745-x9mlp\" (UID: \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.256352 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6tg4\" (UniqueName: \"kubernetes.io/projected/ec75d180-568e-42d8-bb4c-4f82d3fa444b-kube-api-access-b6tg4\") pod \"dnsmasq-dns-758b79db4c-ccck9\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.256375 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-config\") pod \"dnsmasq-dns-758b79db4c-ccck9\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.256394 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-dns-svc\") pod \"dnsmasq-dns-758b79db4c-ccck9\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.257201 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-dns-svc\") pod \"dnsmasq-dns-758b79db4c-ccck9\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.257833 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfe01f2f-670f-4c70-8139-f2988dd1bda9-config\") pod \"dnsmasq-dns-7bfcb9d745-x9mlp\" (UID: \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.258138 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-config\") pod \"dnsmasq-dns-758b79db4c-ccck9\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.280978 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfqlz\" (UniqueName: \"kubernetes.io/projected/bfe01f2f-670f-4c70-8139-f2988dd1bda9-kube-api-access-wfqlz\") pod \"dnsmasq-dns-7bfcb9d745-x9mlp\" (UID: \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\") " pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.287714 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6tg4\" (UniqueName: \"kubernetes.io/projected/ec75d180-568e-42d8-bb4c-4f82d3fa444b-kube-api-access-b6tg4\") pod \"dnsmasq-dns-758b79db4c-ccck9\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.292780 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.371591 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.809400 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-x9mlp"] Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.822492 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 10:20:34 crc kubenswrapper[4923]: I1009 10:20:34.916148 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ccck9"] Oct 09 10:20:35 crc kubenswrapper[4923]: I1009 10:20:35.095378 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" event={"ID":"bfe01f2f-670f-4c70-8139-f2988dd1bda9","Type":"ContainerStarted","Data":"b4f268b5d64fb07cf8a98e29aed172d1b2cbd6a5b4ab311cc6dd67f04413c114"} Oct 09 10:20:35 crc kubenswrapper[4923]: I1009 10:20:35.097063 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-ccck9" event={"ID":"ec75d180-568e-42d8-bb4c-4f82d3fa444b","Type":"ContainerStarted","Data":"9888c8a318ce904e559ce6fc983ba580df70657b09b223ed12adae39237027a7"} Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.094811 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-x9mlp"] Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.122887 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-644597f84c-5znns"] Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.124197 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.141560 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-5znns"] Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.213606 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwk28\" (UniqueName: \"kubernetes.io/projected/636a4a5e-828e-4c70-933a-85f0e69f0dcc-kube-api-access-rwk28\") pod \"dnsmasq-dns-644597f84c-5znns\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.217380 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-dns-svc\") pod \"dnsmasq-dns-644597f84c-5znns\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.217584 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-config\") pod \"dnsmasq-dns-644597f84c-5znns\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.319570 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-dns-svc\") pod \"dnsmasq-dns-644597f84c-5znns\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.319675 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-config\") pod \"dnsmasq-dns-644597f84c-5znns\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.319698 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwk28\" (UniqueName: \"kubernetes.io/projected/636a4a5e-828e-4c70-933a-85f0e69f0dcc-kube-api-access-rwk28\") pod \"dnsmasq-dns-644597f84c-5znns\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.320568 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-config\") pod \"dnsmasq-dns-644597f84c-5znns\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.321181 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-dns-svc\") pod \"dnsmasq-dns-644597f84c-5znns\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.373994 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwk28\" (UniqueName: \"kubernetes.io/projected/636a4a5e-828e-4c70-933a-85f0e69f0dcc-kube-api-access-rwk28\") pod \"dnsmasq-dns-644597f84c-5znns\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.386696 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ccck9"] Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.409559 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-77597f887-r8x99"] Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.411388 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.459585 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.494860 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-r8x99"] Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.526141 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2fwvn\" (UniqueName: \"kubernetes.io/projected/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-kube-api-access-2fwvn\") pod \"dnsmasq-dns-77597f887-r8x99\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.526266 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-config\") pod \"dnsmasq-dns-77597f887-r8x99\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.526291 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-dns-svc\") pod \"dnsmasq-dns-77597f887-r8x99\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.630518 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-dns-svc\") pod \"dnsmasq-dns-77597f887-r8x99\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.630587 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2fwvn\" (UniqueName: \"kubernetes.io/projected/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-kube-api-access-2fwvn\") pod \"dnsmasq-dns-77597f887-r8x99\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.630690 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-config\") pod \"dnsmasq-dns-77597f887-r8x99\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.631681 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-config\") pod \"dnsmasq-dns-77597f887-r8x99\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.631719 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-dns-svc\") pod \"dnsmasq-dns-77597f887-r8x99\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.675816 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2fwvn\" (UniqueName: \"kubernetes.io/projected/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-kube-api-access-2fwvn\") pod \"dnsmasq-dns-77597f887-r8x99\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:37 crc kubenswrapper[4923]: I1009 10:20:37.762914 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.114301 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-5znns"] Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.264040 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.265590 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.267933 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.268082 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.268247 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.268327 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-frbbl" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.268516 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.268543 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.270131 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-77597f887-r8x99"] Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.274097 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.288609 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.344558 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.344895 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.344935 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.344958 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8kw9\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-kube-api-access-h8kw9\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.345004 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/deeda18d-93f0-47c0-8cdd-358505802a0d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.345034 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-config-data\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.345063 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.345092 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/deeda18d-93f0-47c0-8cdd-358505802a0d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.345119 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.345170 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.345199 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446717 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446779 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446826 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446844 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446869 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446885 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8kw9\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-kube-api-access-h8kw9\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446916 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/deeda18d-93f0-47c0-8cdd-358505802a0d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446941 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-config-data\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446961 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446982 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/deeda18d-93f0-47c0-8cdd-358505802a0d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.446999 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.447115 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.447429 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.447837 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.448905 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-config-data\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.449991 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.450167 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-server-conf\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.456896 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.456933 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.460515 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/deeda18d-93f0-47c0-8cdd-358505802a0d-pod-info\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.462339 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/deeda18d-93f0-47c0-8cdd-358505802a0d-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.464641 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8kw9\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-kube-api-access-h8kw9\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.482998 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.604468 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.673735 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.675396 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.681183 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.681579 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.681731 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.682067 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g4rg7" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.682114 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.682080 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.682197 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.702558 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.751572 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.751620 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.751641 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29358129-c9b5-4aca-bd00-0473171b3e96-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.751666 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.751849 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.751887 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29358129-c9b5-4aca-bd00-0473171b3e96-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.751909 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.751946 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgv9h\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-kube-api-access-cgv9h\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.752034 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.752079 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.752113 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.854631 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.854702 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.854729 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.854773 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29358129-c9b5-4aca-bd00-0473171b3e96-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.854801 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.854845 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.854876 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29358129-c9b5-4aca-bd00-0473171b3e96-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.854904 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.854949 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgv9h\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-kube-api-access-cgv9h\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.855003 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.855035 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.855239 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.855370 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.855840 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.856664 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.856813 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.857046 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.859426 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.860120 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29358129-c9b5-4aca-bd00-0473171b3e96-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.862782 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.873657 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgv9h\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-kube-api-access-cgv9h\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.892269 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29358129-c9b5-4aca-bd00-0473171b3e96-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:38 crc kubenswrapper[4923]: I1009 10:20:38.892372 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:39 crc kubenswrapper[4923]: I1009 10:20:39.002207 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.164426 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.167316 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.170895 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.171268 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.171968 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-25r75" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.172053 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.172599 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.181614 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.199686 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.280952 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.281030 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.281049 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-secrets\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.281100 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bk5n\" (UniqueName: \"kubernetes.io/projected/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-kube-api-access-5bk5n\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.281168 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-config-data-default\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.281319 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.281367 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-kolla-config\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.281478 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.281542 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.383185 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.383261 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.383283 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-secrets\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.383309 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bk5n\" (UniqueName: \"kubernetes.io/projected/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-kube-api-access-5bk5n\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.383329 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-config-data-default\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.383656 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.383685 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-kolla-config\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.383713 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.383731 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.384051 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.384295 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-config-data-generated\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.384952 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-config-data-default\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.385498 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-kolla-config\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.386236 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-operator-scripts\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.405589 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.405686 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.405913 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-secrets\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.409209 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bk5n\" (UniqueName: \"kubernetes.io/projected/3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d-kube-api-access-5bk5n\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.426975 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"openstack-galera-0\" (UID: \"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d\") " pod="openstack/openstack-galera-0" Oct 09 10:20:40 crc kubenswrapper[4923]: I1009 10:20:40.492076 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.133744 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.135859 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.138203 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.139044 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-f2qb8" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.139703 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.142359 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.156092 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.174575 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-5znns" event={"ID":"636a4a5e-828e-4c70-933a-85f0e69f0dcc","Type":"ContainerStarted","Data":"0d39ab5fc1983fb5e93e6e9e2ce0a634a77d0220899592c8716ed523a89c17ac"} Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.300827 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.300927 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.301001 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.301085 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.301153 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.301196 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dgj9\" (UniqueName: \"kubernetes.io/projected/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-kube-api-access-5dgj9\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.301224 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.301290 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.301324 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.402869 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.402928 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.402967 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dgj9\" (UniqueName: \"kubernetes.io/projected/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-kube-api-access-5dgj9\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.402995 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.403735 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.403794 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.403801 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.403822 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.404259 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.404408 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.404603 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.404667 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.405089 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.411787 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.414483 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.414733 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.414975 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.427595 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dgj9\" (UniqueName: \"kubernetes.io/projected/4d9b867d-d4b6-4fe5-ad0b-4b5909577d25-kube-api-access-5dgj9\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.438116 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25\") " pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.469682 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.550867 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.552295 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.555438 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.555464 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.555849 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-7pnwl" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.569663 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.608062 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c68f8faa-8010-49a5-949a-94933174149b-config-data\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.608143 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c68f8faa-8010-49a5-949a-94933174149b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.608179 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c68f8faa-8010-49a5-949a-94933174149b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.608363 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9p5cz\" (UniqueName: \"kubernetes.io/projected/c68f8faa-8010-49a5-949a-94933174149b-kube-api-access-9p5cz\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.608397 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c68f8faa-8010-49a5-949a-94933174149b-kolla-config\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.710175 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9p5cz\" (UniqueName: \"kubernetes.io/projected/c68f8faa-8010-49a5-949a-94933174149b-kube-api-access-9p5cz\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.710219 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c68f8faa-8010-49a5-949a-94933174149b-kolla-config\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.710315 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c68f8faa-8010-49a5-949a-94933174149b-config-data\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.710355 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c68f8faa-8010-49a5-949a-94933174149b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.710374 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c68f8faa-8010-49a5-949a-94933174149b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.711953 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c68f8faa-8010-49a5-949a-94933174149b-kolla-config\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.712859 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c68f8faa-8010-49a5-949a-94933174149b-config-data\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.714814 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c68f8faa-8010-49a5-949a-94933174149b-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.715563 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c68f8faa-8010-49a5-949a-94933174149b-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.729368 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9p5cz\" (UniqueName: \"kubernetes.io/projected/c68f8faa-8010-49a5-949a-94933174149b-kube-api-access-9p5cz\") pod \"memcached-0\" (UID: \"c68f8faa-8010-49a5-949a-94933174149b\") " pod="openstack/memcached-0" Oct 09 10:20:41 crc kubenswrapper[4923]: I1009 10:20:41.876414 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 09 10:20:43 crc kubenswrapper[4923]: I1009 10:20:43.288191 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 10:20:43 crc kubenswrapper[4923]: I1009 10:20:43.289509 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 10:20:43 crc kubenswrapper[4923]: I1009 10:20:43.300561 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-cpmzk" Oct 09 10:20:43 crc kubenswrapper[4923]: I1009 10:20:43.311978 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 10:20:43 crc kubenswrapper[4923]: I1009 10:20:43.439113 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxs2d\" (UniqueName: \"kubernetes.io/projected/c94c527a-e960-451f-87a8-4ed58af17475-kube-api-access-bxs2d\") pod \"kube-state-metrics-0\" (UID: \"c94c527a-e960-451f-87a8-4ed58af17475\") " pod="openstack/kube-state-metrics-0" Oct 09 10:20:43 crc kubenswrapper[4923]: I1009 10:20:43.540931 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxs2d\" (UniqueName: \"kubernetes.io/projected/c94c527a-e960-451f-87a8-4ed58af17475-kube-api-access-bxs2d\") pod \"kube-state-metrics-0\" (UID: \"c94c527a-e960-451f-87a8-4ed58af17475\") " pod="openstack/kube-state-metrics-0" Oct 09 10:20:43 crc kubenswrapper[4923]: I1009 10:20:43.572951 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxs2d\" (UniqueName: \"kubernetes.io/projected/c94c527a-e960-451f-87a8-4ed58af17475-kube-api-access-bxs2d\") pod \"kube-state-metrics-0\" (UID: \"c94c527a-e960-451f-87a8-4ed58af17475\") " pod="openstack/kube-state-metrics-0" Oct 09 10:20:43 crc kubenswrapper[4923]: I1009 10:20:43.614458 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.238202 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-r8x99" event={"ID":"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c","Type":"ContainerStarted","Data":"d00d3f4242ee89c9260c5a866af5f003b74cbf1cfc065e589349959d2372a2e6"} Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.739135 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jgs4j"] Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.740792 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.742494 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.743362 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-6lwgs" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.747077 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.750155 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-m429c"] Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.752964 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.755170 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgs4j"] Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.797445 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blcp7\" (UniqueName: \"kubernetes.io/projected/4c323704-d81e-4c9b-a58d-b9943c2f55d7-kube-api-access-blcp7\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.797595 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4c323704-d81e-4c9b-a58d-b9943c2f55d7-var-run-ovn\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.797626 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c323704-d81e-4c9b-a58d-b9943c2f55d7-scripts\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.797655 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4c323704-d81e-4c9b-a58d-b9943c2f55d7-var-log-ovn\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.797679 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-var-lib\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.797711 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-var-log\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.797821 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c323704-d81e-4c9b-a58d-b9943c2f55d7-combined-ca-bundle\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.797845 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-var-run\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.797874 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd7nc\" (UniqueName: \"kubernetes.io/projected/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-kube-api-access-wd7nc\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.798194 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c323704-d81e-4c9b-a58d-b9943c2f55d7-ovn-controller-tls-certs\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.798256 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-scripts\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.798291 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4c323704-d81e-4c9b-a58d-b9943c2f55d7-var-run\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.798316 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-etc-ovs\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.801687 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-m429c"] Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900007 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4c323704-d81e-4c9b-a58d-b9943c2f55d7-var-log-ovn\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900095 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-var-lib\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900139 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-var-log\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900187 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c323704-d81e-4c9b-a58d-b9943c2f55d7-combined-ca-bundle\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900217 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-var-run\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900243 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd7nc\" (UniqueName: \"kubernetes.io/projected/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-kube-api-access-wd7nc\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900280 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c323704-d81e-4c9b-a58d-b9943c2f55d7-ovn-controller-tls-certs\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900319 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-scripts\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900367 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4c323704-d81e-4c9b-a58d-b9943c2f55d7-var-run\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900399 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-etc-ovs\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900475 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blcp7\" (UniqueName: \"kubernetes.io/projected/4c323704-d81e-4c9b-a58d-b9943c2f55d7-kube-api-access-blcp7\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900506 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c323704-d81e-4c9b-a58d-b9943c2f55d7-scripts\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.900529 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4c323704-d81e-4c9b-a58d-b9943c2f55d7-var-run-ovn\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.901276 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4c323704-d81e-4c9b-a58d-b9943c2f55d7-var-run-ovn\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.901450 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4c323704-d81e-4c9b-a58d-b9943c2f55d7-var-log-ovn\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.901603 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-var-lib\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.901736 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-var-log\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.905510 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-scripts\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.906889 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-var-run\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.907149 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-etc-ovs\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.907414 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4c323704-d81e-4c9b-a58d-b9943c2f55d7-var-run\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.916769 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c323704-d81e-4c9b-a58d-b9943c2f55d7-ovn-controller-tls-certs\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.916843 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4c323704-d81e-4c9b-a58d-b9943c2f55d7-scripts\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.920822 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd7nc\" (UniqueName: \"kubernetes.io/projected/1d1e3e9d-552f-441c-ae67-8af6e4fda12b-kube-api-access-wd7nc\") pod \"ovn-controller-ovs-m429c\" (UID: \"1d1e3e9d-552f-441c-ae67-8af6e4fda12b\") " pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.925160 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blcp7\" (UniqueName: \"kubernetes.io/projected/4c323704-d81e-4c9b-a58d-b9943c2f55d7-kube-api-access-blcp7\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:46 crc kubenswrapper[4923]: I1009 10:20:46.935158 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c323704-d81e-4c9b-a58d-b9943c2f55d7-combined-ca-bundle\") pod \"ovn-controller-jgs4j\" (UID: \"4c323704-d81e-4c9b-a58d-b9943c2f55d7\") " pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:47 crc kubenswrapper[4923]: I1009 10:20:47.071720 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgs4j" Oct 09 10:20:47 crc kubenswrapper[4923]: I1009 10:20:47.097166 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:20:48 crc kubenswrapper[4923]: I1009 10:20:48.953453 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 09 10:20:48 crc kubenswrapper[4923]: I1009 10:20:48.956347 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:48 crc kubenswrapper[4923]: I1009 10:20:48.959233 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-hch9k" Oct 09 10:20:48 crc kubenswrapper[4923]: I1009 10:20:48.959259 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 09 10:20:48 crc kubenswrapper[4923]: I1009 10:20:48.959233 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 09 10:20:48 crc kubenswrapper[4923]: I1009 10:20:48.959248 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 09 10:20:48 crc kubenswrapper[4923]: I1009 10:20:48.959918 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 09 10:20:48 crc kubenswrapper[4923]: I1009 10:20:48.973890 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.038636 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.038995 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/da46f3c8-e318-4f39-94e8-814eb2d775bc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.039199 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/da46f3c8-e318-4f39-94e8-814eb2d775bc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.039342 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czv59\" (UniqueName: \"kubernetes.io/projected/da46f3c8-e318-4f39-94e8-814eb2d775bc-kube-api-access-czv59\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.039435 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da46f3c8-e318-4f39-94e8-814eb2d775bc-config\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.039487 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da46f3c8-e318-4f39-94e8-814eb2d775bc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.039522 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da46f3c8-e318-4f39-94e8-814eb2d775bc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.039627 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da46f3c8-e318-4f39-94e8-814eb2d775bc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.141217 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/da46f3c8-e318-4f39-94e8-814eb2d775bc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.141285 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czv59\" (UniqueName: \"kubernetes.io/projected/da46f3c8-e318-4f39-94e8-814eb2d775bc-kube-api-access-czv59\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.141314 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da46f3c8-e318-4f39-94e8-814eb2d775bc-config\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.141341 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da46f3c8-e318-4f39-94e8-814eb2d775bc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.141364 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da46f3c8-e318-4f39-94e8-814eb2d775bc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.141409 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da46f3c8-e318-4f39-94e8-814eb2d775bc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.141454 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.141486 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/da46f3c8-e318-4f39-94e8-814eb2d775bc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.141896 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/da46f3c8-e318-4f39-94e8-814eb2d775bc-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.142537 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.143015 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/da46f3c8-e318-4f39-94e8-814eb2d775bc-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.143687 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da46f3c8-e318-4f39-94e8-814eb2d775bc-config\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.149702 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/da46f3c8-e318-4f39-94e8-814eb2d775bc-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.150693 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da46f3c8-e318-4f39-94e8-814eb2d775bc-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.151988 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da46f3c8-e318-4f39-94e8-814eb2d775bc-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.169610 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.177780 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czv59\" (UniqueName: \"kubernetes.io/projected/da46f3c8-e318-4f39-94e8-814eb2d775bc-kube-api-access-czv59\") pod \"ovsdbserver-nb-0\" (UID: \"da46f3c8-e318-4f39-94e8-814eb2d775bc\") " pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:49 crc kubenswrapper[4923]: I1009 10:20:49.296794 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.916781 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.945391 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.948601 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.949623 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.949974 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.956136 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-tmn96" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.970497 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.971669 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6sxs\" (UniqueName: \"kubernetes.io/projected/46f24e5a-a778-40c2-837e-e5cfda1992ff-kube-api-access-b6sxs\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.971713 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f24e5a-a778-40c2-837e-e5cfda1992ff-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.971736 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f24e5a-a778-40c2-837e-e5cfda1992ff-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.971843 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.971869 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46f24e5a-a778-40c2-837e-e5cfda1992ff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.971886 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f24e5a-a778-40c2-837e-e5cfda1992ff-config\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.971923 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/46f24e5a-a778-40c2-837e-e5cfda1992ff-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:50 crc kubenswrapper[4923]: I1009 10:20:50.971942 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f24e5a-a778-40c2-837e-e5cfda1992ff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.073353 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.073406 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46f24e5a-a778-40c2-837e-e5cfda1992ff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.073422 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f24e5a-a778-40c2-837e-e5cfda1992ff-config\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.073458 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/46f24e5a-a778-40c2-837e-e5cfda1992ff-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.073481 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f24e5a-a778-40c2-837e-e5cfda1992ff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.073520 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6sxs\" (UniqueName: \"kubernetes.io/projected/46f24e5a-a778-40c2-837e-e5cfda1992ff-kube-api-access-b6sxs\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.073543 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f24e5a-a778-40c2-837e-e5cfda1992ff-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.073563 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f24e5a-a778-40c2-837e-e5cfda1992ff-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.073654 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.074631 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/46f24e5a-a778-40c2-837e-e5cfda1992ff-config\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.076968 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/46f24e5a-a778-40c2-837e-e5cfda1992ff-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.077867 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/46f24e5a-a778-40c2-837e-e5cfda1992ff-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.080338 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f24e5a-a778-40c2-837e-e5cfda1992ff-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.081068 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f24e5a-a778-40c2-837e-e5cfda1992ff-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.093843 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f24e5a-a778-40c2-837e-e5cfda1992ff-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.098084 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6sxs\" (UniqueName: \"kubernetes.io/projected/46f24e5a-a778-40c2-837e-e5cfda1992ff-kube-api-access-b6sxs\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.100108 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"46f24e5a-a778-40c2-837e-e5cfda1992ff\") " pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:51 crc kubenswrapper[4923]: I1009 10:20:51.275852 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 09 10:20:52 crc kubenswrapper[4923]: I1009 10:20:52.376291 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 09 10:20:53 crc kubenswrapper[4923]: W1009 10:20:53.048574 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ecb3847_ac13_4e9a_b8db_abafa3ca6a0d.slice/crio-bf8711226fb2eeac87da73139c24771df2af662ad6d37c9d01d3cd22f64f8a47 WatchSource:0}: Error finding container bf8711226fb2eeac87da73139c24771df2af662ad6d37c9d01d3cd22f64f8a47: Status 404 returned error can't find the container with id bf8711226fb2eeac87da73139c24771df2af662ad6d37c9d01d3cd22f64f8a47 Oct 09 10:20:53 crc kubenswrapper[4923]: E1009 10:20:53.050222 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 09 10:20:53 crc kubenswrapper[4923]: E1009 10:20:53.050411 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wfqlz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7bfcb9d745-x9mlp_openstack(bfe01f2f-670f-4c70-8139-f2988dd1bda9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:20:53 crc kubenswrapper[4923]: E1009 10:20:53.051853 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" podUID="bfe01f2f-670f-4c70-8139-f2988dd1bda9" Oct 09 10:20:53 crc kubenswrapper[4923]: E1009 10:20:53.105132 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df" Oct 09 10:20:53 crc kubenswrapper[4923]: E1009 10:20:53.105392 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b6tg4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-758b79db4c-ccck9_openstack(ec75d180-568e-42d8-bb4c-4f82d3fa444b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:20:53 crc kubenswrapper[4923]: E1009 10:20:53.106918 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-758b79db4c-ccck9" podUID="ec75d180-568e-42d8-bb4c-4f82d3fa444b" Oct 09 10:20:53 crc kubenswrapper[4923]: I1009 10:20:53.326133 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d","Type":"ContainerStarted","Data":"bf8711226fb2eeac87da73139c24771df2af662ad6d37c9d01d3cd22f64f8a47"} Oct 09 10:20:53 crc kubenswrapper[4923]: I1009 10:20:53.489465 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 10:20:53 crc kubenswrapper[4923]: I1009 10:20:53.705278 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 09 10:20:53 crc kubenswrapper[4923]: I1009 10:20:53.741685 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 10:20:53 crc kubenswrapper[4923]: W1009 10:20:53.756255 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddeeda18d_93f0_47c0_8cdd_358505802a0d.slice/crio-af175916751b1f0164efb54579af39beb3b4e2d64eca40acb107ac4195578305 WatchSource:0}: Error finding container af175916751b1f0164efb54579af39beb3b4e2d64eca40acb107ac4195578305: Status 404 returned error can't find the container with id af175916751b1f0164efb54579af39beb3b4e2d64eca40acb107ac4195578305 Oct 09 10:20:53 crc kubenswrapper[4923]: I1009 10:20:53.869311 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:53 crc kubenswrapper[4923]: I1009 10:20:53.876411 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.034589 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6tg4\" (UniqueName: \"kubernetes.io/projected/ec75d180-568e-42d8-bb4c-4f82d3fa444b-kube-api-access-b6tg4\") pod \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.034634 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-config\") pod \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.034657 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-dns-svc\") pod \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\" (UID: \"ec75d180-568e-42d8-bb4c-4f82d3fa444b\") " Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.034696 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfqlz\" (UniqueName: \"kubernetes.io/projected/bfe01f2f-670f-4c70-8139-f2988dd1bda9-kube-api-access-wfqlz\") pod \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\" (UID: \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\") " Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.034806 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfe01f2f-670f-4c70-8139-f2988dd1bda9-config\") pod \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\" (UID: \"bfe01f2f-670f-4c70-8139-f2988dd1bda9\") " Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.035655 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bfe01f2f-670f-4c70-8139-f2988dd1bda9-config" (OuterVolumeSpecName: "config") pod "bfe01f2f-670f-4c70-8139-f2988dd1bda9" (UID: "bfe01f2f-670f-4c70-8139-f2988dd1bda9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.036269 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ec75d180-568e-42d8-bb4c-4f82d3fa444b" (UID: "ec75d180-568e-42d8-bb4c-4f82d3fa444b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.036480 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-config" (OuterVolumeSpecName: "config") pod "ec75d180-568e-42d8-bb4c-4f82d3fa444b" (UID: "ec75d180-568e-42d8-bb4c-4f82d3fa444b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.041666 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfe01f2f-670f-4c70-8139-f2988dd1bda9-kube-api-access-wfqlz" (OuterVolumeSpecName: "kube-api-access-wfqlz") pod "bfe01f2f-670f-4c70-8139-f2988dd1bda9" (UID: "bfe01f2f-670f-4c70-8139-f2988dd1bda9"). InnerVolumeSpecName "kube-api-access-wfqlz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.042020 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec75d180-568e-42d8-bb4c-4f82d3fa444b-kube-api-access-b6tg4" (OuterVolumeSpecName: "kube-api-access-b6tg4") pod "ec75d180-568e-42d8-bb4c-4f82d3fa444b" (UID: "ec75d180-568e-42d8-bb4c-4f82d3fa444b"). InnerVolumeSpecName "kube-api-access-b6tg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.137202 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6tg4\" (UniqueName: \"kubernetes.io/projected/ec75d180-568e-42d8-bb4c-4f82d3fa444b-kube-api-access-b6tg4\") on node \"crc\" DevicePath \"\"" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.137245 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.137259 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ec75d180-568e-42d8-bb4c-4f82d3fa444b-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.137271 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfqlz\" (UniqueName: \"kubernetes.io/projected/bfe01f2f-670f-4c70-8139-f2988dd1bda9-kube-api-access-wfqlz\") on node \"crc\" DevicePath \"\"" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.137285 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfe01f2f-670f-4c70-8139-f2988dd1bda9-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.227488 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.389491 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c68f8faa-8010-49a5-949a-94933174149b","Type":"ContainerStarted","Data":"808dad4d233ce874cc63c6bee85e94bd52a7e87b77d2f07ee2983c87874ae967"} Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.407414 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" event={"ID":"bfe01f2f-670f-4c70-8139-f2988dd1bda9","Type":"ContainerDied","Data":"b4f268b5d64fb07cf8a98e29aed172d1b2cbd6a5b4ab311cc6dd67f04413c114"} Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.407520 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bfcb9d745-x9mlp" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.422854 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-758b79db4c-ccck9" event={"ID":"ec75d180-568e-42d8-bb4c-4f82d3fa444b","Type":"ContainerDied","Data":"9888c8a318ce904e559ce6fc983ba580df70657b09b223ed12adae39237027a7"} Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.423014 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-758b79db4c-ccck9" Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.448503 4923 generic.go:334] "Generic (PLEG): container finished" podID="da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" containerID="fb6c4540dfb31517999d4e0ddeac1e633209bc176a5ef02ee20cb8be88938348" exitCode=0 Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.448618 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-r8x99" event={"ID":"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c","Type":"ContainerDied","Data":"fb6c4540dfb31517999d4e0ddeac1e633209bc176a5ef02ee20cb8be88938348"} Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.475186 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c94c527a-e960-451f-87a8-4ed58af17475","Type":"ContainerStarted","Data":"68cdc099aaf0cae2a0d152b4c401d7ad8ca27f0fe82f793ac3e2e11b741c1723"} Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.479354 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.488002 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29358129-c9b5-4aca-bd00-0473171b3e96","Type":"ContainerStarted","Data":"294c65561254a0bddd82c5a3b05d485133d49c5ae0f5468cc593fa6993694739"} Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.490251 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"deeda18d-93f0-47c0-8cdd-358505802a0d","Type":"ContainerStarted","Data":"af175916751b1f0164efb54579af39beb3b4e2d64eca40acb107ac4195578305"} Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.493992 4923 generic.go:334] "Generic (PLEG): container finished" podID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" containerID="d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67" exitCode=0 Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.494048 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-5znns" event={"ID":"636a4a5e-828e-4c70-933a-85f0e69f0dcc","Type":"ContainerDied","Data":"d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67"} Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.496559 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgs4j"] Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.572377 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.597478 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-x9mlp"] Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.602197 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.602244 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:20:54 crc kubenswrapper[4923]: W1009 10:20:54.605252 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46f24e5a_a778_40c2_837e_e5cfda1992ff.slice/crio-e01cfd14ee65df734c7a398de320e8360c6a9fee6fb56910c2796c62ad25d851 WatchSource:0}: Error finding container e01cfd14ee65df734c7a398de320e8360c6a9fee6fb56910c2796c62ad25d851: Status 404 returned error can't find the container with id e01cfd14ee65df734c7a398de320e8360c6a9fee6fb56910c2796c62ad25d851 Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.617127 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bfcb9d745-x9mlp"] Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.641828 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ccck9"] Oct 09 10:20:54 crc kubenswrapper[4923]: I1009 10:20:54.645616 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-758b79db4c-ccck9"] Oct 09 10:20:54 crc kubenswrapper[4923]: E1009 10:20:54.804373 4923 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Oct 09 10:20:54 crc kubenswrapper[4923]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/636a4a5e-828e-4c70-933a-85f0e69f0dcc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 09 10:20:54 crc kubenswrapper[4923]: > podSandboxID="0d39ab5fc1983fb5e93e6e9e2ce0a634a77d0220899592c8716ed523a89c17ac" Oct 09 10:20:54 crc kubenswrapper[4923]: E1009 10:20:54.805075 4923 kuberuntime_manager.go:1274] "Unhandled Error" err=< Oct 09 10:20:54 crc kubenswrapper[4923]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:c4e71b2158fd939dad8b8e705273493051d3023273d23b279f2699dce6db33df,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rwk28,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-644597f84c-5znns_openstack(636a4a5e-828e-4c70-933a-85f0e69f0dcc): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/636a4a5e-828e-4c70-933a-85f0e69f0dcc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Oct 09 10:20:54 crc kubenswrapper[4923]: > logger="UnhandledError" Oct 09 10:20:54 crc kubenswrapper[4923]: E1009 10:20:54.806785 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/636a4a5e-828e-4c70-933a-85f0e69f0dcc/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-644597f84c-5znns" podUID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" Oct 09 10:20:55 crc kubenswrapper[4923]: I1009 10:20:55.268104 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-m429c"] Oct 09 10:20:55 crc kubenswrapper[4923]: I1009 10:20:55.494336 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 09 10:20:55 crc kubenswrapper[4923]: I1009 10:20:55.523016 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-r8x99" event={"ID":"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c","Type":"ContainerStarted","Data":"84403dbff6e0b896180c9792b00118a8f3732aee0e861bb99fc0a76a4c9d8c21"} Oct 09 10:20:55 crc kubenswrapper[4923]: I1009 10:20:55.523123 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:20:55 crc kubenswrapper[4923]: I1009 10:20:55.525134 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"46f24e5a-a778-40c2-837e-e5cfda1992ff","Type":"ContainerStarted","Data":"e01cfd14ee65df734c7a398de320e8360c6a9fee6fb56910c2796c62ad25d851"} Oct 09 10:20:55 crc kubenswrapper[4923]: I1009 10:20:55.527190 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25","Type":"ContainerStarted","Data":"fd10d737608a9da22b76a347a04c255a0183244b8ca2d9d318804453af23aca5"} Oct 09 10:20:55 crc kubenswrapper[4923]: I1009 10:20:55.529858 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgs4j" event={"ID":"4c323704-d81e-4c9b-a58d-b9943c2f55d7","Type":"ContainerStarted","Data":"02a419fbf99db323588b4c6546dfd1377f48e98bd021fab4a368ba34ffaa3c4e"} Oct 09 10:20:55 crc kubenswrapper[4923]: I1009 10:20:55.548649 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-77597f887-r8x99" podStartSLOduration=10.953801098 podStartE2EDuration="18.548600852s" podCreationTimestamp="2025-10-09 10:20:37 +0000 UTC" firstStartedPulling="2025-10-09 10:20:45.819721596 +0000 UTC m=+931.887903352" lastFinishedPulling="2025-10-09 10:20:53.41452135 +0000 UTC m=+939.482703106" observedRunningTime="2025-10-09 10:20:55.543465079 +0000 UTC m=+941.611646845" watchObservedRunningTime="2025-10-09 10:20:55.548600852 +0000 UTC m=+941.616782608" Oct 09 10:20:55 crc kubenswrapper[4923]: W1009 10:20:55.813315 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda46f3c8_e318_4f39_94e8_814eb2d775bc.slice/crio-26d6e7baefe8d5890bb75252a87bf2081170d31f0d995ea8aed9b6578f30a29b WatchSource:0}: Error finding container 26d6e7baefe8d5890bb75252a87bf2081170d31f0d995ea8aed9b6578f30a29b: Status 404 returned error can't find the container with id 26d6e7baefe8d5890bb75252a87bf2081170d31f0d995ea8aed9b6578f30a29b Oct 09 10:20:55 crc kubenswrapper[4923]: W1009 10:20:55.854269 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d1e3e9d_552f_441c_ae67_8af6e4fda12b.slice/crio-59aca18f1f3a098da1aedb7ddf8efd3677cd7db1d27154256733564a74428ce3 WatchSource:0}: Error finding container 59aca18f1f3a098da1aedb7ddf8efd3677cd7db1d27154256733564a74428ce3: Status 404 returned error can't find the container with id 59aca18f1f3a098da1aedb7ddf8efd3677cd7db1d27154256733564a74428ce3 Oct 09 10:20:56 crc kubenswrapper[4923]: I1009 10:20:56.537839 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m429c" event={"ID":"1d1e3e9d-552f-441c-ae67-8af6e4fda12b","Type":"ContainerStarted","Data":"59aca18f1f3a098da1aedb7ddf8efd3677cd7db1d27154256733564a74428ce3"} Oct 09 10:20:56 crc kubenswrapper[4923]: I1009 10:20:56.539435 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"da46f3c8-e318-4f39-94e8-814eb2d775bc","Type":"ContainerStarted","Data":"26d6e7baefe8d5890bb75252a87bf2081170d31f0d995ea8aed9b6578f30a29b"} Oct 09 10:20:56 crc kubenswrapper[4923]: I1009 10:20:56.616375 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfe01f2f-670f-4c70-8139-f2988dd1bda9" path="/var/lib/kubelet/pods/bfe01f2f-670f-4c70-8139-f2988dd1bda9/volumes" Oct 09 10:20:56 crc kubenswrapper[4923]: I1009 10:20:56.616945 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec75d180-568e-42d8-bb4c-4f82d3fa444b" path="/var/lib/kubelet/pods/ec75d180-568e-42d8-bb4c-4f82d3fa444b/volumes" Oct 09 10:21:02 crc kubenswrapper[4923]: I1009 10:21:02.766049 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:21:02 crc kubenswrapper[4923]: I1009 10:21:02.827360 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-5znns"] Oct 09 10:21:04 crc kubenswrapper[4923]: I1009 10:21:04.610421 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-644597f84c-5znns" podUID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" containerName="dnsmasq-dns" containerID="cri-o://c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08" gracePeriod=10 Oct 09 10:21:04 crc kubenswrapper[4923]: I1009 10:21:04.614663 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-5znns" event={"ID":"636a4a5e-828e-4c70-933a-85f0e69f0dcc","Type":"ContainerStarted","Data":"c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08"} Oct 09 10:21:04 crc kubenswrapper[4923]: I1009 10:21:04.614725 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:21:04 crc kubenswrapper[4923]: I1009 10:21:04.614743 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 09 10:21:04 crc kubenswrapper[4923]: I1009 10:21:04.666692 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=14.559551376 podStartE2EDuration="23.666665899s" podCreationTimestamp="2025-10-09 10:20:41 +0000 UTC" firstStartedPulling="2025-10-09 10:20:53.718809159 +0000 UTC m=+939.786990925" lastFinishedPulling="2025-10-09 10:21:02.825923692 +0000 UTC m=+948.894105448" observedRunningTime="2025-10-09 10:21:04.659739247 +0000 UTC m=+950.727921023" watchObservedRunningTime="2025-10-09 10:21:04.666665899 +0000 UTC m=+950.734847655" Oct 09 10:21:04 crc kubenswrapper[4923]: I1009 10:21:04.684088 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-644597f84c-5znns" podStartSLOduration=15.541918415 podStartE2EDuration="27.684065872s" podCreationTimestamp="2025-10-09 10:20:37 +0000 UTC" firstStartedPulling="2025-10-09 10:20:41.090528738 +0000 UTC m=+927.158710504" lastFinishedPulling="2025-10-09 10:20:53.232676205 +0000 UTC m=+939.300857961" observedRunningTime="2025-10-09 10:21:04.680431541 +0000 UTC m=+950.748613297" watchObservedRunningTime="2025-10-09 10:21:04.684065872 +0000 UTC m=+950.752247628" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.351073 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.391599 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwk28\" (UniqueName: \"kubernetes.io/projected/636a4a5e-828e-4c70-933a-85f0e69f0dcc-kube-api-access-rwk28\") pod \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.392015 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-dns-svc\") pod \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.392120 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-config\") pod \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\" (UID: \"636a4a5e-828e-4c70-933a-85f0e69f0dcc\") " Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.422087 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636a4a5e-828e-4c70-933a-85f0e69f0dcc-kube-api-access-rwk28" (OuterVolumeSpecName: "kube-api-access-rwk28") pod "636a4a5e-828e-4c70-933a-85f0e69f0dcc" (UID: "636a4a5e-828e-4c70-933a-85f0e69f0dcc"). InnerVolumeSpecName "kube-api-access-rwk28". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.449173 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "636a4a5e-828e-4c70-933a-85f0e69f0dcc" (UID: "636a4a5e-828e-4c70-933a-85f0e69f0dcc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.454322 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-config" (OuterVolumeSpecName: "config") pod "636a4a5e-828e-4c70-933a-85f0e69f0dcc" (UID: "636a4a5e-828e-4c70-933a-85f0e69f0dcc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.493887 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.493927 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/636a4a5e-828e-4c70-933a-85f0e69f0dcc-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.493945 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwk28\" (UniqueName: \"kubernetes.io/projected/636a4a5e-828e-4c70-933a-85f0e69f0dcc-kube-api-access-rwk28\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.623032 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"da46f3c8-e318-4f39-94e8-814eb2d775bc","Type":"ContainerStarted","Data":"22627c4297bcad2c74c02f94269d3e2bf7c6b3d253b2955d5f9f546c5214aad7"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.624708 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29358129-c9b5-4aca-bd00-0473171b3e96","Type":"ContainerStarted","Data":"863db0b0cc594cea2e1a5f5b6284f95afd669dcec8c8ece965c04e9aa52c6f22"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.626117 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"deeda18d-93f0-47c0-8cdd-358505802a0d","Type":"ContainerStarted","Data":"2049d58a742c0f3521d999b5b6d1678f1c3179e14b2ad167f60ebabd76e38a1f"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.627904 4923 generic.go:334] "Generic (PLEG): container finished" podID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" containerID="c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08" exitCode=0 Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.627937 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-5znns" event={"ID":"636a4a5e-828e-4c70-933a-85f0e69f0dcc","Type":"ContainerDied","Data":"c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.627968 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-644597f84c-5znns" event={"ID":"636a4a5e-828e-4c70-933a-85f0e69f0dcc","Type":"ContainerDied","Data":"0d39ab5fc1983fb5e93e6e9e2ce0a634a77d0220899592c8716ed523a89c17ac"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.627990 4923 scope.go:117] "RemoveContainer" containerID="c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.628169 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-644597f84c-5znns" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.629263 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m429c" event={"ID":"1d1e3e9d-552f-441c-ae67-8af6e4fda12b","Type":"ContainerStarted","Data":"63a6eee488ce3bc5d88a443e9d156e019f42a67a63682a4aed4ad76f326127f4"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.631386 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgs4j" event={"ID":"4c323704-d81e-4c9b-a58d-b9943c2f55d7","Type":"ContainerStarted","Data":"19b162a8e6294ec8662b1e988e18c2235c124f7604eb3bfd12b52db1d1e03335"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.631460 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-jgs4j" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.632713 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c94c527a-e960-451f-87a8-4ed58af17475","Type":"ContainerStarted","Data":"e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.632832 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.634195 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d","Type":"ContainerStarted","Data":"5f07d6290930f9b67687545605c9603a6a76dbe9f8022d4a855f27a75afe8880"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.635870 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"46f24e5a-a778-40c2-837e-e5cfda1992ff","Type":"ContainerStarted","Data":"467ed4849e8b63ad68c1fb1a084e1fe064237b42e2d9bfebfdae68a22a7723f3"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.637219 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25","Type":"ContainerStarted","Data":"b31087fe55defb32f15058dd25d8d6aac5c60bcf3f6372e369cedea0da82714f"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.638487 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c68f8faa-8010-49a5-949a-94933174149b","Type":"ContainerStarted","Data":"81b5dfdd00e74042c38a6ff0d49aa91b8f0a233517b4751164142760c0e44bcd"} Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.647564 4923 scope.go:117] "RemoveContainer" containerID="d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.685303 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-5znns"] Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.711802 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-644597f84c-5znns"] Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.714719 4923 scope.go:117] "RemoveContainer" containerID="c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08" Oct 09 10:21:05 crc kubenswrapper[4923]: E1009 10:21:05.715563 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08\": container with ID starting with c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08 not found: ID does not exist" containerID="c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.715611 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08"} err="failed to get container status \"c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08\": rpc error: code = NotFound desc = could not find container \"c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08\": container with ID starting with c6bdef5d3da87c7637f3d5c16177f827dc6db68c39a5138fe3dd614f68c12a08 not found: ID does not exist" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.715646 4923 scope.go:117] "RemoveContainer" containerID="d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67" Oct 09 10:21:05 crc kubenswrapper[4923]: E1009 10:21:05.716167 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67\": container with ID starting with d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67 not found: ID does not exist" containerID="d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.716199 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67"} err="failed to get container status \"d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67\": rpc error: code = NotFound desc = could not find container \"d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67\": container with ID starting with d9333b7c380aae446e1cd2e3c40f63a25216a3f9b91ef9e76bdeb1b3ed5b3a67 not found: ID does not exist" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.756716 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-jgs4j" podStartSLOduration=10.912699711 podStartE2EDuration="19.756695649s" podCreationTimestamp="2025-10-09 10:20:46 +0000 UTC" firstStartedPulling="2025-10-09 10:20:54.471386919 +0000 UTC m=+940.539568675" lastFinishedPulling="2025-10-09 10:21:03.315382857 +0000 UTC m=+949.383564613" observedRunningTime="2025-10-09 10:21:05.751013361 +0000 UTC m=+951.819195127" watchObservedRunningTime="2025-10-09 10:21:05.756695649 +0000 UTC m=+951.824877405" Oct 09 10:21:05 crc kubenswrapper[4923]: I1009 10:21:05.812245 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=13.270383144 podStartE2EDuration="22.812228062s" podCreationTimestamp="2025-10-09 10:20:43 +0000 UTC" firstStartedPulling="2025-10-09 10:20:54.328077324 +0000 UTC m=+940.396259080" lastFinishedPulling="2025-10-09 10:21:03.869922242 +0000 UTC m=+949.938103998" observedRunningTime="2025-10-09 10:21:05.812114669 +0000 UTC m=+951.880296435" watchObservedRunningTime="2025-10-09 10:21:05.812228062 +0000 UTC m=+951.880409818" Oct 09 10:21:06 crc kubenswrapper[4923]: I1009 10:21:06.614696 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" path="/var/lib/kubelet/pods/636a4a5e-828e-4c70-933a-85f0e69f0dcc/volumes" Oct 09 10:21:06 crc kubenswrapper[4923]: I1009 10:21:06.653410 4923 generic.go:334] "Generic (PLEG): container finished" podID="1d1e3e9d-552f-441c-ae67-8af6e4fda12b" containerID="63a6eee488ce3bc5d88a443e9d156e019f42a67a63682a4aed4ad76f326127f4" exitCode=0 Oct 09 10:21:06 crc kubenswrapper[4923]: I1009 10:21:06.654694 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m429c" event={"ID":"1d1e3e9d-552f-441c-ae67-8af6e4fda12b","Type":"ContainerDied","Data":"63a6eee488ce3bc5d88a443e9d156e019f42a67a63682a4aed4ad76f326127f4"} Oct 09 10:21:07 crc kubenswrapper[4923]: I1009 10:21:07.669541 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m429c" event={"ID":"1d1e3e9d-552f-441c-ae67-8af6e4fda12b","Type":"ContainerStarted","Data":"413d4ea50cef18f6bf8b507d0ae86944434352b874169f3344e9b098cf757a10"} Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.674039 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-t8h9w"] Oct 09 10:21:08 crc kubenswrapper[4923]: E1009 10:21:08.674685 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" containerName="dnsmasq-dns" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.674697 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" containerName="dnsmasq-dns" Oct 09 10:21:08 crc kubenswrapper[4923]: E1009 10:21:08.674711 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" containerName="init" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.674717 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" containerName="init" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.674960 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="636a4a5e-828e-4c70-933a-85f0e69f0dcc" containerName="dnsmasq-dns" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.675576 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.679224 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.692689 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-t8h9w"] Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.746427 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ce6ae725-8786-4973-bafe-8d6548ee51bb-ovs-rundir\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.746520 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7hww\" (UniqueName: \"kubernetes.io/projected/ce6ae725-8786-4973-bafe-8d6548ee51bb-kube-api-access-c7hww\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.746599 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce6ae725-8786-4973-bafe-8d6548ee51bb-combined-ca-bundle\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.746644 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6ae725-8786-4973-bafe-8d6548ee51bb-config\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.746676 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce6ae725-8786-4973-bafe-8d6548ee51bb-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.746702 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ce6ae725-8786-4973-bafe-8d6548ee51bb-ovn-rundir\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.828718 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-s6zh9"] Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.830519 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.833443 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.848577 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce6ae725-8786-4973-bafe-8d6548ee51bb-combined-ca-bundle\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.848649 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6ae725-8786-4973-bafe-8d6548ee51bb-config\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.848694 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce6ae725-8786-4973-bafe-8d6548ee51bb-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.848728 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ce6ae725-8786-4973-bafe-8d6548ee51bb-ovn-rundir\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.848786 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ce6ae725-8786-4973-bafe-8d6548ee51bb-ovs-rundir\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.848836 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7hww\" (UniqueName: \"kubernetes.io/projected/ce6ae725-8786-4973-bafe-8d6548ee51bb-kube-api-access-c7hww\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.852550 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-s6zh9"] Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.854548 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/ce6ae725-8786-4973-bafe-8d6548ee51bb-ovn-rundir\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.854612 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/ce6ae725-8786-4973-bafe-8d6548ee51bb-ovs-rundir\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.854849 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce6ae725-8786-4973-bafe-8d6548ee51bb-config\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.870609 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce6ae725-8786-4973-bafe-8d6548ee51bb-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.870702 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7hww\" (UniqueName: \"kubernetes.io/projected/ce6ae725-8786-4973-bafe-8d6548ee51bb-kube-api-access-c7hww\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.879442 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce6ae725-8786-4973-bafe-8d6548ee51bb-combined-ca-bundle\") pod \"ovn-controller-metrics-t8h9w\" (UID: \"ce6ae725-8786-4973-bafe-8d6548ee51bb\") " pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.953229 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dn49\" (UniqueName: \"kubernetes.io/projected/7d72169f-dc10-4721-a880-9325170da82e-kube-api-access-8dn49\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.953305 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-ovsdbserver-nb\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.953342 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-dns-svc\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.953384 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-config\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:08 crc kubenswrapper[4923]: I1009 10:21:08.999434 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-s6zh9"] Oct 09 10:21:09 crc kubenswrapper[4923]: E1009 10:21:09.010292 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-8dn49 ovsdbserver-nb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" podUID="7d72169f-dc10-4721-a880-9325170da82e" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.048899 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-wtm94"] Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.050587 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.055439 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dn49\" (UniqueName: \"kubernetes.io/projected/7d72169f-dc10-4721-a880-9325170da82e-kube-api-access-8dn49\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.055539 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-ovsdbserver-nb\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.055592 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-dns-svc\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.055647 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-config\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.056669 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-ovsdbserver-nb\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.056883 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-dns-svc\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.057238 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-config\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.066279 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.070314 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-t8h9w" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.086500 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-wtm94"] Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.121429 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dn49\" (UniqueName: \"kubernetes.io/projected/7d72169f-dc10-4721-a880-9325170da82e-kube-api-access-8dn49\") pod \"dnsmasq-dns-f6b595d95-s6zh9\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.161042 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.161107 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.161143 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7c5h\" (UniqueName: \"kubernetes.io/projected/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-kube-api-access-t7c5h\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.161190 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.161280 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-config\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.264207 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.264765 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.264813 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7c5h\" (UniqueName: \"kubernetes.io/projected/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-kube-api-access-t7c5h\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.264869 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.264969 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-config\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.265766 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-nb\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.266056 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-sb\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.270833 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-dns-svc\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.270846 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-config\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.284313 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7c5h\" (UniqueName: \"kubernetes.io/projected/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-kube-api-access-t7c5h\") pod \"dnsmasq-dns-dc9d58d7-wtm94\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.379968 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.605264 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-t8h9w"] Oct 09 10:21:09 crc kubenswrapper[4923]: W1009 10:21:09.610283 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podce6ae725_8786_4973_bafe_8d6548ee51bb.slice/crio-9d7d764823e781a6edbc8d40511e910c953d18193783522daf4e80fc241e3f4c WatchSource:0}: Error finding container 9d7d764823e781a6edbc8d40511e910c953d18193783522daf4e80fc241e3f4c: Status 404 returned error can't find the container with id 9d7d764823e781a6edbc8d40511e910c953d18193783522daf4e80fc241e3f4c Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.688213 4923 generic.go:334] "Generic (PLEG): container finished" podID="3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d" containerID="5f07d6290930f9b67687545605c9603a6a76dbe9f8022d4a855f27a75afe8880" exitCode=0 Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.688328 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d","Type":"ContainerDied","Data":"5f07d6290930f9b67687545605c9603a6a76dbe9f8022d4a855f27a75afe8880"} Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.702930 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"da46f3c8-e318-4f39-94e8-814eb2d775bc","Type":"ContainerStarted","Data":"32ba8057f6520972882bdb697c025bdf91f45b61d3c7f63b6e830793261a5900"} Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.704943 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"46f24e5a-a778-40c2-837e-e5cfda1992ff","Type":"ContainerStarted","Data":"f3cda4b55e4c9800ccae30e9f30cbbc15b3e2fda56345fe77a8bc89a886016b5"} Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.710674 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-t8h9w" event={"ID":"ce6ae725-8786-4973-bafe-8d6548ee51bb","Type":"ContainerStarted","Data":"9d7d764823e781a6edbc8d40511e910c953d18193783522daf4e80fc241e3f4c"} Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.713927 4923 generic.go:334] "Generic (PLEG): container finished" podID="4d9b867d-d4b6-4fe5-ad0b-4b5909577d25" containerID="b31087fe55defb32f15058dd25d8d6aac5c60bcf3f6372e369cedea0da82714f" exitCode=0 Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.714021 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25","Type":"ContainerDied","Data":"b31087fe55defb32f15058dd25d8d6aac5c60bcf3f6372e369cedea0da82714f"} Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.717743 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.717921 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-m429c" event={"ID":"1d1e3e9d-552f-441c-ae67-8af6e4fda12b","Type":"ContainerStarted","Data":"88a22888ff1b498013a17846b01c73c28042a25459bbcbc3c61316e00f0a18da"} Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.718089 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.718131 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.737736 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=6.860951914 podStartE2EDuration="20.737708919s" podCreationTimestamp="2025-10-09 10:20:49 +0000 UTC" firstStartedPulling="2025-10-09 10:20:54.620566155 +0000 UTC m=+940.688747901" lastFinishedPulling="2025-10-09 10:21:08.49732314 +0000 UTC m=+954.565504906" observedRunningTime="2025-10-09 10:21:09.730093838 +0000 UTC m=+955.798275594" watchObservedRunningTime="2025-10-09 10:21:09.737708919 +0000 UTC m=+955.805890675" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.743207 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.757347 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=10.072327398 podStartE2EDuration="22.757318575s" podCreationTimestamp="2025-10-09 10:20:47 +0000 UTC" firstStartedPulling="2025-10-09 10:20:55.815838411 +0000 UTC m=+941.884020167" lastFinishedPulling="2025-10-09 10:21:08.500829578 +0000 UTC m=+954.569011344" observedRunningTime="2025-10-09 10:21:09.751933385 +0000 UTC m=+955.820115141" watchObservedRunningTime="2025-10-09 10:21:09.757318575 +0000 UTC m=+955.825500331" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.813462 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-m429c" podStartSLOduration=16.364545667 podStartE2EDuration="23.813444715s" podCreationTimestamp="2025-10-09 10:20:46 +0000 UTC" firstStartedPulling="2025-10-09 10:20:55.860097181 +0000 UTC m=+941.928278927" lastFinishedPulling="2025-10-09 10:21:03.308996219 +0000 UTC m=+949.377177975" observedRunningTime="2025-10-09 10:21:09.810476012 +0000 UTC m=+955.878657768" watchObservedRunningTime="2025-10-09 10:21:09.813444715 +0000 UTC m=+955.881626471" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.884740 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-ovsdbserver-nb\") pod \"7d72169f-dc10-4721-a880-9325170da82e\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.884891 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dn49\" (UniqueName: \"kubernetes.io/projected/7d72169f-dc10-4721-a880-9325170da82e-kube-api-access-8dn49\") pod \"7d72169f-dc10-4721-a880-9325170da82e\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.884957 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-dns-svc\") pod \"7d72169f-dc10-4721-a880-9325170da82e\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.884996 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-config\") pod \"7d72169f-dc10-4721-a880-9325170da82e\" (UID: \"7d72169f-dc10-4721-a880-9325170da82e\") " Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.886102 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7d72169f-dc10-4721-a880-9325170da82e" (UID: "7d72169f-dc10-4721-a880-9325170da82e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.886132 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-config" (OuterVolumeSpecName: "config") pod "7d72169f-dc10-4721-a880-9325170da82e" (UID: "7d72169f-dc10-4721-a880-9325170da82e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.886593 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7d72169f-dc10-4721-a880-9325170da82e" (UID: "7d72169f-dc10-4721-a880-9325170da82e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.893280 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d72169f-dc10-4721-a880-9325170da82e-kube-api-access-8dn49" (OuterVolumeSpecName: "kube-api-access-8dn49") pod "7d72169f-dc10-4721-a880-9325170da82e" (UID: "7d72169f-dc10-4721-a880-9325170da82e"). InnerVolumeSpecName "kube-api-access-8dn49". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.930736 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-wtm94"] Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.987212 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.987265 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.987280 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d72169f-dc10-4721-a880-9325170da82e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:09 crc kubenswrapper[4923]: I1009 10:21:09.987293 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dn49\" (UniqueName: \"kubernetes.io/projected/7d72169f-dc10-4721-a880-9325170da82e-kube-api-access-8dn49\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.297352 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.342508 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.731188 4923 generic.go:334] "Generic (PLEG): container finished" podID="1eb424ba-b604-47d9-88f0-c268ccdd5aa3" containerID="03e098255c35400936fd2b183de45fe3394363fcff61b95395f96f6875f29384" exitCode=0 Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.731328 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" event={"ID":"1eb424ba-b604-47d9-88f0-c268ccdd5aa3","Type":"ContainerDied","Data":"03e098255c35400936fd2b183de45fe3394363fcff61b95395f96f6875f29384"} Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.731561 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" event={"ID":"1eb424ba-b604-47d9-88f0-c268ccdd5aa3","Type":"ContainerStarted","Data":"5a8d2a2c0846a6f4d3fcd38566c524b52550797ba733600f61cbc9e04807ac30"} Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.734089 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-t8h9w" event={"ID":"ce6ae725-8786-4973-bafe-8d6548ee51bb","Type":"ContainerStarted","Data":"56a5925ca3fbdbdb22799d2cb79edb08524cafb3e00dea5741c23648b96b46c9"} Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.737351 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"4d9b867d-d4b6-4fe5-ad0b-4b5909577d25","Type":"ContainerStarted","Data":"7965090be6219ad0a7122cb34e584ed55d03c25a55d69e92760b23df3cdb12be"} Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.740401 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d","Type":"ContainerStarted","Data":"26e74911fcbdce09945d8fae6d2bc1c68979e650a7a549a2f4e633e29084bad3"} Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.740684 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6b595d95-s6zh9" Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.741947 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.790805 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-t8h9w" podStartSLOduration=2.790778902 podStartE2EDuration="2.790778902s" podCreationTimestamp="2025-10-09 10:21:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:21:10.774543711 +0000 UTC m=+956.842725467" watchObservedRunningTime="2025-10-09 10:21:10.790778902 +0000 UTC m=+956.858960678" Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.811134 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=22.005626692 podStartE2EDuration="31.811111057s" podCreationTimestamp="2025-10-09 10:20:39 +0000 UTC" firstStartedPulling="2025-10-09 10:20:53.052860157 +0000 UTC m=+939.121041913" lastFinishedPulling="2025-10-09 10:21:02.858344522 +0000 UTC m=+948.926526278" observedRunningTime="2025-10-09 10:21:10.804666868 +0000 UTC m=+956.872848644" watchObservedRunningTime="2025-10-09 10:21:10.811111057 +0000 UTC m=+956.879292823" Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.816567 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.841225 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=21.49228687 podStartE2EDuration="30.841201354s" podCreationTimestamp="2025-10-09 10:20:40 +0000 UTC" firstStartedPulling="2025-10-09 10:20:54.38835925 +0000 UTC m=+940.456541006" lastFinishedPulling="2025-10-09 10:21:03.737273734 +0000 UTC m=+949.805455490" observedRunningTime="2025-10-09 10:21:10.839220888 +0000 UTC m=+956.907402644" watchObservedRunningTime="2025-10-09 10:21:10.841201354 +0000 UTC m=+956.909383110" Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.892268 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-s6zh9"] Oct 09 10:21:10 crc kubenswrapper[4923]: I1009 10:21:10.896823 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f6b595d95-s6zh9"] Oct 09 10:21:11 crc kubenswrapper[4923]: I1009 10:21:11.277239 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 09 10:21:11 crc kubenswrapper[4923]: I1009 10:21:11.469900 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 09 10:21:11 crc kubenswrapper[4923]: I1009 10:21:11.470672 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 09 10:21:11 crc kubenswrapper[4923]: I1009 10:21:11.753875 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" event={"ID":"1eb424ba-b604-47d9-88f0-c268ccdd5aa3","Type":"ContainerStarted","Data":"53e977ac3d7dba4288d2e45bacf549b66fee42a44884f8b0403e44dfc8108f39"} Oct 09 10:21:11 crc kubenswrapper[4923]: I1009 10:21:11.755976 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:11 crc kubenswrapper[4923]: I1009 10:21:11.786942 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" podStartSLOduration=2.786911832 podStartE2EDuration="2.786911832s" podCreationTimestamp="2025-10-09 10:21:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:21:11.77639337 +0000 UTC m=+957.844575166" watchObservedRunningTime="2025-10-09 10:21:11.786911832 +0000 UTC m=+957.855093608" Oct 09 10:21:11 crc kubenswrapper[4923]: I1009 10:21:11.878991 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 09 10:21:12 crc kubenswrapper[4923]: I1009 10:21:12.277415 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 09 10:21:12 crc kubenswrapper[4923]: I1009 10:21:12.313987 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 09 10:21:12 crc kubenswrapper[4923]: I1009 10:21:12.612377 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d72169f-dc10-4721-a880-9325170da82e" path="/var/lib/kubelet/pods/7d72169f-dc10-4721-a880-9325170da82e/volumes" Oct 09 10:21:12 crc kubenswrapper[4923]: I1009 10:21:12.806364 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.033374 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.034670 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.038090 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.038498 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-7gt4k" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.038690 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.038914 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.053823 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ef458c9-e8d0-4251-924b-c46ef4cb203b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.053880 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ef458c9-e8d0-4251-924b-c46ef4cb203b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.053912 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9ef458c9-e8d0-4251-924b-c46ef4cb203b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.053929 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbw6b\" (UniqueName: \"kubernetes.io/projected/9ef458c9-e8d0-4251-924b-c46ef4cb203b-kube-api-access-qbw6b\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.054019 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef458c9-e8d0-4251-924b-c46ef4cb203b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.054098 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ef458c9-e8d0-4251-924b-c46ef4cb203b-config\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.054120 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ef458c9-e8d0-4251-924b-c46ef4cb203b-scripts\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.069659 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.155886 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef458c9-e8d0-4251-924b-c46ef4cb203b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.155969 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ef458c9-e8d0-4251-924b-c46ef4cb203b-config\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.155992 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ef458c9-e8d0-4251-924b-c46ef4cb203b-scripts\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.156037 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ef458c9-e8d0-4251-924b-c46ef4cb203b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.156066 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ef458c9-e8d0-4251-924b-c46ef4cb203b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.156094 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9ef458c9-e8d0-4251-924b-c46ef4cb203b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.156126 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbw6b\" (UniqueName: \"kubernetes.io/projected/9ef458c9-e8d0-4251-924b-c46ef4cb203b-kube-api-access-qbw6b\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.156627 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9ef458c9-e8d0-4251-924b-c46ef4cb203b-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.157138 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ef458c9-e8d0-4251-924b-c46ef4cb203b-config\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.157256 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9ef458c9-e8d0-4251-924b-c46ef4cb203b-scripts\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.162888 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ef458c9-e8d0-4251-924b-c46ef4cb203b-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.162914 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ef458c9-e8d0-4251-924b-c46ef4cb203b-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.170916 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef458c9-e8d0-4251-924b-c46ef4cb203b-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.176594 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbw6b\" (UniqueName: \"kubernetes.io/projected/9ef458c9-e8d0-4251-924b-c46ef4cb203b-kube-api-access-qbw6b\") pod \"ovn-northd-0\" (UID: \"9ef458c9-e8d0-4251-924b-c46ef4cb203b\") " pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.353937 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.627334 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 09 10:21:13 crc kubenswrapper[4923]: I1009 10:21:13.824085 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 09 10:21:14 crc kubenswrapper[4923]: I1009 10:21:14.780209 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9ef458c9-e8d0-4251-924b-c46ef4cb203b","Type":"ContainerStarted","Data":"6235dcc36d39f79d55a6bfa063cfdc06f7d701f4ee9e4df972600125689820b8"} Oct 09 10:21:15 crc kubenswrapper[4923]: E1009 10:21:15.060260 4923 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.120:59676->38.102.83.120:45195: read tcp 38.102.83.120:59676->38.102.83.120:45195: read: connection reset by peer Oct 09 10:21:15 crc kubenswrapper[4923]: I1009 10:21:15.795608 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9ef458c9-e8d0-4251-924b-c46ef4cb203b","Type":"ContainerStarted","Data":"ef40deceff33b4daa0232c1975c04b7fba340156fb14cd317b77d36a1ae60a59"} Oct 09 10:21:15 crc kubenswrapper[4923]: I1009 10:21:15.796212 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 09 10:21:15 crc kubenswrapper[4923]: I1009 10:21:15.796231 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9ef458c9-e8d0-4251-924b-c46ef4cb203b","Type":"ContainerStarted","Data":"65a4b9c99c99bcc7c2c5c0690f310cbaabd6835ec9cb99963999ac6e4f221b26"} Oct 09 10:21:15 crc kubenswrapper[4923]: I1009 10:21:15.848770 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.8749212910000002 podStartE2EDuration="2.848725449s" podCreationTimestamp="2025-10-09 10:21:13 +0000 UTC" firstStartedPulling="2025-10-09 10:21:13.835161008 +0000 UTC m=+959.903342764" lastFinishedPulling="2025-10-09 10:21:14.808965166 +0000 UTC m=+960.877146922" observedRunningTime="2025-10-09 10:21:15.828393093 +0000 UTC m=+961.896574859" watchObservedRunningTime="2025-10-09 10:21:15.848725449 +0000 UTC m=+961.916907215" Oct 09 10:21:17 crc kubenswrapper[4923]: I1009 10:21:17.761610 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 09 10:21:17 crc kubenswrapper[4923]: I1009 10:21:17.851199 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="4d9b867d-d4b6-4fe5-ad0b-4b5909577d25" containerName="galera" probeResult="failure" output=< Oct 09 10:21:17 crc kubenswrapper[4923]: wsrep_local_state_comment (Joined) differs from Synced Oct 09 10:21:17 crc kubenswrapper[4923]: > Oct 09 10:21:19 crc kubenswrapper[4923]: I1009 10:21:19.382032 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:21:19 crc kubenswrapper[4923]: I1009 10:21:19.445250 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-r8x99"] Oct 09 10:21:19 crc kubenswrapper[4923]: I1009 10:21:19.445843 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-77597f887-r8x99" podUID="da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" containerName="dnsmasq-dns" containerID="cri-o://84403dbff6e0b896180c9792b00118a8f3732aee0e861bb99fc0a76a4c9d8c21" gracePeriod=10 Oct 09 10:21:19 crc kubenswrapper[4923]: I1009 10:21:19.840239 4923 generic.go:334] "Generic (PLEG): container finished" podID="da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" containerID="84403dbff6e0b896180c9792b00118a8f3732aee0e861bb99fc0a76a4c9d8c21" exitCode=0 Oct 09 10:21:19 crc kubenswrapper[4923]: I1009 10:21:19.840306 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-r8x99" event={"ID":"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c","Type":"ContainerDied","Data":"84403dbff6e0b896180c9792b00118a8f3732aee0e861bb99fc0a76a4c9d8c21"} Oct 09 10:21:20 crc kubenswrapper[4923]: I1009 10:21:20.493355 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 09 10:21:20 crc kubenswrapper[4923]: I1009 10:21:20.493924 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 09 10:21:20 crc kubenswrapper[4923]: I1009 10:21:20.557395 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 09 10:21:20 crc kubenswrapper[4923]: I1009 10:21:20.898329 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.535118 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.601390 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-vxmdr"] Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.607949 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vxmdr" Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.613345 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vxmdr"] Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.708688 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tjsw\" (UniqueName: \"kubernetes.io/projected/4e8ce207-1c54-43ea-8f35-a1a47088d249-kube-api-access-2tjsw\") pod \"keystone-db-create-vxmdr\" (UID: \"4e8ce207-1c54-43ea-8f35-a1a47088d249\") " pod="openstack/keystone-db-create-vxmdr" Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.771822 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-679ld"] Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.773867 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-679ld" Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.779992 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-679ld"] Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.811277 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tjsw\" (UniqueName: \"kubernetes.io/projected/4e8ce207-1c54-43ea-8f35-a1a47088d249-kube-api-access-2tjsw\") pod \"keystone-db-create-vxmdr\" (UID: \"4e8ce207-1c54-43ea-8f35-a1a47088d249\") " pod="openstack/keystone-db-create-vxmdr" Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.846210 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tjsw\" (UniqueName: \"kubernetes.io/projected/4e8ce207-1c54-43ea-8f35-a1a47088d249-kube-api-access-2tjsw\") pod \"keystone-db-create-vxmdr\" (UID: \"4e8ce207-1c54-43ea-8f35-a1a47088d249\") " pod="openstack/keystone-db-create-vxmdr" Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.913719 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdndc\" (UniqueName: \"kubernetes.io/projected/e1f95ef9-7942-49c4-971b-9871c3a84b37-kube-api-access-qdndc\") pod \"placement-db-create-679ld\" (UID: \"e1f95ef9-7942-49c4-971b-9871c3a84b37\") " pod="openstack/placement-db-create-679ld" Oct 09 10:21:21 crc kubenswrapper[4923]: I1009 10:21:21.932291 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vxmdr" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.016949 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdndc\" (UniqueName: \"kubernetes.io/projected/e1f95ef9-7942-49c4-971b-9871c3a84b37-kube-api-access-qdndc\") pod \"placement-db-create-679ld\" (UID: \"e1f95ef9-7942-49c4-971b-9871c3a84b37\") " pod="openstack/placement-db-create-679ld" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.023814 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-sbbl8"] Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.024977 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sbbl8" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.029332 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sbbl8"] Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.039380 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdndc\" (UniqueName: \"kubernetes.io/projected/e1f95ef9-7942-49c4-971b-9871c3a84b37-kube-api-access-qdndc\") pod \"placement-db-create-679ld\" (UID: \"e1f95ef9-7942-49c4-971b-9871c3a84b37\") " pod="openstack/placement-db-create-679ld" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.086682 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.095791 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-679ld" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.121543 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkssz\" (UniqueName: \"kubernetes.io/projected/d3b07ecd-e8d3-494a-aefe-5c8204428bb6-kube-api-access-wkssz\") pod \"glance-db-create-sbbl8\" (UID: \"d3b07ecd-e8d3-494a-aefe-5c8204428bb6\") " pod="openstack/glance-db-create-sbbl8" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.223255 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-config\") pod \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.223387 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-dns-svc\") pod \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.223410 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2fwvn\" (UniqueName: \"kubernetes.io/projected/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-kube-api-access-2fwvn\") pod \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\" (UID: \"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c\") " Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.223647 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkssz\" (UniqueName: \"kubernetes.io/projected/d3b07ecd-e8d3-494a-aefe-5c8204428bb6-kube-api-access-wkssz\") pod \"glance-db-create-sbbl8\" (UID: \"d3b07ecd-e8d3-494a-aefe-5c8204428bb6\") " pod="openstack/glance-db-create-sbbl8" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.229390 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-kube-api-access-2fwvn" (OuterVolumeSpecName: "kube-api-access-2fwvn") pod "da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" (UID: "da6a1f7d-7a40-4d25-b5a6-58518bf3a35c"). InnerVolumeSpecName "kube-api-access-2fwvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.246718 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkssz\" (UniqueName: \"kubernetes.io/projected/d3b07ecd-e8d3-494a-aefe-5c8204428bb6-kube-api-access-wkssz\") pod \"glance-db-create-sbbl8\" (UID: \"d3b07ecd-e8d3-494a-aefe-5c8204428bb6\") " pod="openstack/glance-db-create-sbbl8" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.266856 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" (UID: "da6a1f7d-7a40-4d25-b5a6-58518bf3a35c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.274928 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-config" (OuterVolumeSpecName: "config") pod "da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" (UID: "da6a1f7d-7a40-4d25-b5a6-58518bf3a35c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.325972 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.326018 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.326029 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2fwvn\" (UniqueName: \"kubernetes.io/projected/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c-kube-api-access-2fwvn\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.394859 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sbbl8" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.405027 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-vxmdr"] Oct 09 10:21:22 crc kubenswrapper[4923]: W1009 10:21:22.410126 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4e8ce207_1c54_43ea_8f35_a1a47088d249.slice/crio-e3a8f213fb2cd0644ae80a886a29714dd4542f676fd4d8f72dea6fc0deca7f50 WatchSource:0}: Error finding container e3a8f213fb2cd0644ae80a886a29714dd4542f676fd4d8f72dea6fc0deca7f50: Status 404 returned error can't find the container with id e3a8f213fb2cd0644ae80a886a29714dd4542f676fd4d8f72dea6fc0deca7f50 Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.563114 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-679ld"] Oct 09 10:21:22 crc kubenswrapper[4923]: W1009 10:21:22.569348 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1f95ef9_7942_49c4_971b_9871c3a84b37.slice/crio-5d55215be2ffeaa94880538597800a04a010501f8066aa50c19a9143ac03c987 WatchSource:0}: Error finding container 5d55215be2ffeaa94880538597800a04a010501f8066aa50c19a9143ac03c987: Status 404 returned error can't find the container with id 5d55215be2ffeaa94880538597800a04a010501f8066aa50c19a9143ac03c987 Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.846735 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-sbbl8"] Oct 09 10:21:22 crc kubenswrapper[4923]: W1009 10:21:22.870267 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd3b07ecd_e8d3_494a_aefe_5c8204428bb6.slice/crio-222dadd1ab825c74683643ebe36ed955287e8166a050853199f37cbc70ee81c4 WatchSource:0}: Error finding container 222dadd1ab825c74683643ebe36ed955287e8166a050853199f37cbc70ee81c4: Status 404 returned error can't find the container with id 222dadd1ab825c74683643ebe36ed955287e8166a050853199f37cbc70ee81c4 Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.872385 4923 generic.go:334] "Generic (PLEG): container finished" podID="e1f95ef9-7942-49c4-971b-9871c3a84b37" containerID="9c70a17e35e220ec215f760024df38d7563a480010a73842a8a9e2b8ad6ea8c7" exitCode=0 Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.872531 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-679ld" event={"ID":"e1f95ef9-7942-49c4-971b-9871c3a84b37","Type":"ContainerDied","Data":"9c70a17e35e220ec215f760024df38d7563a480010a73842a8a9e2b8ad6ea8c7"} Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.872606 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-679ld" event={"ID":"e1f95ef9-7942-49c4-971b-9871c3a84b37","Type":"ContainerStarted","Data":"5d55215be2ffeaa94880538597800a04a010501f8066aa50c19a9143ac03c987"} Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.875090 4923 generic.go:334] "Generic (PLEG): container finished" podID="4e8ce207-1c54-43ea-8f35-a1a47088d249" containerID="bd71ad4b8fbe98ab192d364066620e857c91a7e87102664e6412ec7329fa015d" exitCode=0 Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.875202 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vxmdr" event={"ID":"4e8ce207-1c54-43ea-8f35-a1a47088d249","Type":"ContainerDied","Data":"bd71ad4b8fbe98ab192d364066620e857c91a7e87102664e6412ec7329fa015d"} Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.875242 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vxmdr" event={"ID":"4e8ce207-1c54-43ea-8f35-a1a47088d249","Type":"ContainerStarted","Data":"e3a8f213fb2cd0644ae80a886a29714dd4542f676fd4d8f72dea6fc0deca7f50"} Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.879883 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-77597f887-r8x99" event={"ID":"da6a1f7d-7a40-4d25-b5a6-58518bf3a35c","Type":"ContainerDied","Data":"d00d3f4242ee89c9260c5a866af5f003b74cbf1cfc065e589349959d2372a2e6"} Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.879932 4923 scope.go:117] "RemoveContainer" containerID="84403dbff6e0b896180c9792b00118a8f3732aee0e861bb99fc0a76a4c9d8c21" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.880091 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-77597f887-r8x99" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.944811 4923 scope.go:117] "RemoveContainer" containerID="fb6c4540dfb31517999d4e0ddeac1e633209bc176a5ef02ee20cb8be88938348" Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.946827 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-77597f887-r8x99"] Oct 09 10:21:22 crc kubenswrapper[4923]: I1009 10:21:22.955077 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-77597f887-r8x99"] Oct 09 10:21:23 crc kubenswrapper[4923]: I1009 10:21:23.893353 4923 generic.go:334] "Generic (PLEG): container finished" podID="d3b07ecd-e8d3-494a-aefe-5c8204428bb6" containerID="12a8551a0ecabb2686d2b0b788853b7b3fe6e2d501c97bc1bb7ff5b30b3cf0b7" exitCode=0 Oct 09 10:21:23 crc kubenswrapper[4923]: I1009 10:21:23.893408 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sbbl8" event={"ID":"d3b07ecd-e8d3-494a-aefe-5c8204428bb6","Type":"ContainerDied","Data":"12a8551a0ecabb2686d2b0b788853b7b3fe6e2d501c97bc1bb7ff5b30b3cf0b7"} Oct 09 10:21:23 crc kubenswrapper[4923]: I1009 10:21:23.893955 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sbbl8" event={"ID":"d3b07ecd-e8d3-494a-aefe-5c8204428bb6","Type":"ContainerStarted","Data":"222dadd1ab825c74683643ebe36ed955287e8166a050853199f37cbc70ee81c4"} Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.314158 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vxmdr" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.320987 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-679ld" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.474340 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdndc\" (UniqueName: \"kubernetes.io/projected/e1f95ef9-7942-49c4-971b-9871c3a84b37-kube-api-access-qdndc\") pod \"e1f95ef9-7942-49c4-971b-9871c3a84b37\" (UID: \"e1f95ef9-7942-49c4-971b-9871c3a84b37\") " Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.474477 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tjsw\" (UniqueName: \"kubernetes.io/projected/4e8ce207-1c54-43ea-8f35-a1a47088d249-kube-api-access-2tjsw\") pod \"4e8ce207-1c54-43ea-8f35-a1a47088d249\" (UID: \"4e8ce207-1c54-43ea-8f35-a1a47088d249\") " Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.481581 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1f95ef9-7942-49c4-971b-9871c3a84b37-kube-api-access-qdndc" (OuterVolumeSpecName: "kube-api-access-qdndc") pod "e1f95ef9-7942-49c4-971b-9871c3a84b37" (UID: "e1f95ef9-7942-49c4-971b-9871c3a84b37"). InnerVolumeSpecName "kube-api-access-qdndc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.488497 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4e8ce207-1c54-43ea-8f35-a1a47088d249-kube-api-access-2tjsw" (OuterVolumeSpecName: "kube-api-access-2tjsw") pod "4e8ce207-1c54-43ea-8f35-a1a47088d249" (UID: "4e8ce207-1c54-43ea-8f35-a1a47088d249"). InnerVolumeSpecName "kube-api-access-2tjsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.576824 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdndc\" (UniqueName: \"kubernetes.io/projected/e1f95ef9-7942-49c4-971b-9871c3a84b37-kube-api-access-qdndc\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.576882 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tjsw\" (UniqueName: \"kubernetes.io/projected/4e8ce207-1c54-43ea-8f35-a1a47088d249-kube-api-access-2tjsw\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.600054 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.600158 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.600224 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.601245 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0371acf1b4e04ebc9cc4694cbecac34256a8ea0829820a4e57e504509cef2cbf"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.601332 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://0371acf1b4e04ebc9cc4694cbecac34256a8ea0829820a4e57e504509cef2cbf" gracePeriod=600 Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.612262 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" path="/var/lib/kubelet/pods/da6a1f7d-7a40-4d25-b5a6-58518bf3a35c/volumes" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.910148 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-vxmdr" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.910600 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-vxmdr" event={"ID":"4e8ce207-1c54-43ea-8f35-a1a47088d249","Type":"ContainerDied","Data":"e3a8f213fb2cd0644ae80a886a29714dd4542f676fd4d8f72dea6fc0deca7f50"} Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.910655 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3a8f213fb2cd0644ae80a886a29714dd4542f676fd4d8f72dea6fc0deca7f50" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.913265 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="0371acf1b4e04ebc9cc4694cbecac34256a8ea0829820a4e57e504509cef2cbf" exitCode=0 Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.913322 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"0371acf1b4e04ebc9cc4694cbecac34256a8ea0829820a4e57e504509cef2cbf"} Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.913348 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"ff1594437e0fe053768cc52266cdd8b6d71b1a9aae0c236740cf019156ef7f8f"} Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.913367 4923 scope.go:117] "RemoveContainer" containerID="af930c04f923630accaccaa4b93be2a9110fb2ade8cd00a8a79620f54fdb4120" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.920857 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-679ld" Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.922098 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-679ld" event={"ID":"e1f95ef9-7942-49c4-971b-9871c3a84b37","Type":"ContainerDied","Data":"5d55215be2ffeaa94880538597800a04a010501f8066aa50c19a9143ac03c987"} Oct 09 10:21:24 crc kubenswrapper[4923]: I1009 10:21:24.922143 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d55215be2ffeaa94880538597800a04a010501f8066aa50c19a9143ac03c987" Oct 09 10:21:25 crc kubenswrapper[4923]: I1009 10:21:25.208374 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sbbl8" Oct 09 10:21:25 crc kubenswrapper[4923]: I1009 10:21:25.293923 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wkssz\" (UniqueName: \"kubernetes.io/projected/d3b07ecd-e8d3-494a-aefe-5c8204428bb6-kube-api-access-wkssz\") pod \"d3b07ecd-e8d3-494a-aefe-5c8204428bb6\" (UID: \"d3b07ecd-e8d3-494a-aefe-5c8204428bb6\") " Oct 09 10:21:25 crc kubenswrapper[4923]: I1009 10:21:25.300222 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3b07ecd-e8d3-494a-aefe-5c8204428bb6-kube-api-access-wkssz" (OuterVolumeSpecName: "kube-api-access-wkssz") pod "d3b07ecd-e8d3-494a-aefe-5c8204428bb6" (UID: "d3b07ecd-e8d3-494a-aefe-5c8204428bb6"). InnerVolumeSpecName "kube-api-access-wkssz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:25 crc kubenswrapper[4923]: I1009 10:21:25.396152 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wkssz\" (UniqueName: \"kubernetes.io/projected/d3b07ecd-e8d3-494a-aefe-5c8204428bb6-kube-api-access-wkssz\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:25 crc kubenswrapper[4923]: I1009 10:21:25.931319 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-sbbl8" event={"ID":"d3b07ecd-e8d3-494a-aefe-5c8204428bb6","Type":"ContainerDied","Data":"222dadd1ab825c74683643ebe36ed955287e8166a050853199f37cbc70ee81c4"} Oct 09 10:21:25 crc kubenswrapper[4923]: I1009 10:21:25.931354 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-sbbl8" Oct 09 10:21:25 crc kubenswrapper[4923]: I1009 10:21:25.931373 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="222dadd1ab825c74683643ebe36ed955287e8166a050853199f37cbc70ee81c4" Oct 09 10:21:28 crc kubenswrapper[4923]: I1009 10:21:28.422366 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.646258 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7825-account-create-phb7p"] Oct 09 10:21:31 crc kubenswrapper[4923]: E1009 10:21:31.647302 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3b07ecd-e8d3-494a-aefe-5c8204428bb6" containerName="mariadb-database-create" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.647324 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3b07ecd-e8d3-494a-aefe-5c8204428bb6" containerName="mariadb-database-create" Oct 09 10:21:31 crc kubenswrapper[4923]: E1009 10:21:31.647337 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4e8ce207-1c54-43ea-8f35-a1a47088d249" containerName="mariadb-database-create" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.647344 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="4e8ce207-1c54-43ea-8f35-a1a47088d249" containerName="mariadb-database-create" Oct 09 10:21:31 crc kubenswrapper[4923]: E1009 10:21:31.647363 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f95ef9-7942-49c4-971b-9871c3a84b37" containerName="mariadb-database-create" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.647370 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f95ef9-7942-49c4-971b-9871c3a84b37" containerName="mariadb-database-create" Oct 09 10:21:31 crc kubenswrapper[4923]: E1009 10:21:31.647384 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" containerName="init" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.647392 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" containerName="init" Oct 09 10:21:31 crc kubenswrapper[4923]: E1009 10:21:31.647405 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" containerName="dnsmasq-dns" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.647412 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" containerName="dnsmasq-dns" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.647575 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3b07ecd-e8d3-494a-aefe-5c8204428bb6" containerName="mariadb-database-create" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.647592 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1f95ef9-7942-49c4-971b-9871c3a84b37" containerName="mariadb-database-create" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.647610 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="da6a1f7d-7a40-4d25-b5a6-58518bf3a35c" containerName="dnsmasq-dns" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.647625 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="4e8ce207-1c54-43ea-8f35-a1a47088d249" containerName="mariadb-database-create" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.648236 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7825-account-create-phb7p" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.656012 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.656042 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7825-account-create-phb7p"] Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.698951 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g9pz\" (UniqueName: \"kubernetes.io/projected/835679c7-23a5-4347-b57c-eb0d600195b4-kube-api-access-8g9pz\") pod \"keystone-7825-account-create-phb7p\" (UID: \"835679c7-23a5-4347-b57c-eb0d600195b4\") " pod="openstack/keystone-7825-account-create-phb7p" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.803583 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g9pz\" (UniqueName: \"kubernetes.io/projected/835679c7-23a5-4347-b57c-eb0d600195b4-kube-api-access-8g9pz\") pod \"keystone-7825-account-create-phb7p\" (UID: \"835679c7-23a5-4347-b57c-eb0d600195b4\") " pod="openstack/keystone-7825-account-create-phb7p" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.820671 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7555-account-create-rqrzd"] Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.822401 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7555-account-create-rqrzd" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.825401 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.831673 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g9pz\" (UniqueName: \"kubernetes.io/projected/835679c7-23a5-4347-b57c-eb0d600195b4-kube-api-access-8g9pz\") pod \"keystone-7825-account-create-phb7p\" (UID: \"835679c7-23a5-4347-b57c-eb0d600195b4\") " pod="openstack/keystone-7825-account-create-phb7p" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.836635 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7555-account-create-rqrzd"] Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.905718 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxpms\" (UniqueName: \"kubernetes.io/projected/31ff6e52-d758-44a7-801c-7a0a9578157a-kube-api-access-pxpms\") pod \"placement-7555-account-create-rqrzd\" (UID: \"31ff6e52-d758-44a7-801c-7a0a9578157a\") " pod="openstack/placement-7555-account-create-rqrzd" Oct 09 10:21:31 crc kubenswrapper[4923]: I1009 10:21:31.971681 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7825-account-create-phb7p" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.009835 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxpms\" (UniqueName: \"kubernetes.io/projected/31ff6e52-d758-44a7-801c-7a0a9578157a-kube-api-access-pxpms\") pod \"placement-7555-account-create-rqrzd\" (UID: \"31ff6e52-d758-44a7-801c-7a0a9578157a\") " pod="openstack/placement-7555-account-create-rqrzd" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.034408 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxpms\" (UniqueName: \"kubernetes.io/projected/31ff6e52-d758-44a7-801c-7a0a9578157a-kube-api-access-pxpms\") pod \"placement-7555-account-create-rqrzd\" (UID: \"31ff6e52-d758-44a7-801c-7a0a9578157a\") " pod="openstack/placement-7555-account-create-rqrzd" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.178780 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7555-account-create-rqrzd" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.203123 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-be11-account-create-9ms9b"] Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.204523 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be11-account-create-9ms9b" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.206460 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-be11-account-create-9ms9b"] Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.209186 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.322321 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpfsf\" (UniqueName: \"kubernetes.io/projected/e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0-kube-api-access-jpfsf\") pod \"glance-be11-account-create-9ms9b\" (UID: \"e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0\") " pod="openstack/glance-be11-account-create-9ms9b" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.424499 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpfsf\" (UniqueName: \"kubernetes.io/projected/e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0-kube-api-access-jpfsf\") pod \"glance-be11-account-create-9ms9b\" (UID: \"e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0\") " pod="openstack/glance-be11-account-create-9ms9b" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.452426 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpfsf\" (UniqueName: \"kubernetes.io/projected/e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0-kube-api-access-jpfsf\") pod \"glance-be11-account-create-9ms9b\" (UID: \"e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0\") " pod="openstack/glance-be11-account-create-9ms9b" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.478596 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7825-account-create-phb7p"] Oct 09 10:21:32 crc kubenswrapper[4923]: W1009 10:21:32.481177 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod835679c7_23a5_4347_b57c_eb0d600195b4.slice/crio-8b7e59bcfb9fc2925e41ca54d354ba43502677c802e203c7328927e3d424a28f WatchSource:0}: Error finding container 8b7e59bcfb9fc2925e41ca54d354ba43502677c802e203c7328927e3d424a28f: Status 404 returned error can't find the container with id 8b7e59bcfb9fc2925e41ca54d354ba43502677c802e203c7328927e3d424a28f Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.592287 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be11-account-create-9ms9b" Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.643479 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7555-account-create-rqrzd"] Oct 09 10:21:32 crc kubenswrapper[4923]: W1009 10:21:32.652260 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31ff6e52_d758_44a7_801c_7a0a9578157a.slice/crio-52b81aa10441ad6d923d34ff88fc20f68c767dd59e258abf6826bebe8b9a2a65 WatchSource:0}: Error finding container 52b81aa10441ad6d923d34ff88fc20f68c767dd59e258abf6826bebe8b9a2a65: Status 404 returned error can't find the container with id 52b81aa10441ad6d923d34ff88fc20f68c767dd59e258abf6826bebe8b9a2a65 Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.997002 4923 generic.go:334] "Generic (PLEG): container finished" podID="31ff6e52-d758-44a7-801c-7a0a9578157a" containerID="44e68a14bb6ea8b8055bb92d53e0cbc3689a8a32507e8bb1adef365e021e5223" exitCode=0 Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.997098 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7555-account-create-rqrzd" event={"ID":"31ff6e52-d758-44a7-801c-7a0a9578157a","Type":"ContainerDied","Data":"44e68a14bb6ea8b8055bb92d53e0cbc3689a8a32507e8bb1adef365e021e5223"} Oct 09 10:21:32 crc kubenswrapper[4923]: I1009 10:21:32.997480 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7555-account-create-rqrzd" event={"ID":"31ff6e52-d758-44a7-801c-7a0a9578157a","Type":"ContainerStarted","Data":"52b81aa10441ad6d923d34ff88fc20f68c767dd59e258abf6826bebe8b9a2a65"} Oct 09 10:21:33 crc kubenswrapper[4923]: I1009 10:21:33.000162 4923 generic.go:334] "Generic (PLEG): container finished" podID="835679c7-23a5-4347-b57c-eb0d600195b4" containerID="650c63ee4e7e5f6c3a703abd4efeec520b335c2ebd3c443573dac709d2abab09" exitCode=0 Oct 09 10:21:33 crc kubenswrapper[4923]: I1009 10:21:33.000212 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7825-account-create-phb7p" event={"ID":"835679c7-23a5-4347-b57c-eb0d600195b4","Type":"ContainerDied","Data":"650c63ee4e7e5f6c3a703abd4efeec520b335c2ebd3c443573dac709d2abab09"} Oct 09 10:21:33 crc kubenswrapper[4923]: I1009 10:21:33.000425 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7825-account-create-phb7p" event={"ID":"835679c7-23a5-4347-b57c-eb0d600195b4","Type":"ContainerStarted","Data":"8b7e59bcfb9fc2925e41ca54d354ba43502677c802e203c7328927e3d424a28f"} Oct 09 10:21:33 crc kubenswrapper[4923]: I1009 10:21:33.059855 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-be11-account-create-9ms9b"] Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.026800 4923 generic.go:334] "Generic (PLEG): container finished" podID="e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0" containerID="98e77b5646898ba123a7413d3eaaffc4bb39735dc78eb2a2ddeab9b565f39be8" exitCode=0 Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.026952 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be11-account-create-9ms9b" event={"ID":"e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0","Type":"ContainerDied","Data":"98e77b5646898ba123a7413d3eaaffc4bb39735dc78eb2a2ddeab9b565f39be8"} Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.027408 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be11-account-create-9ms9b" event={"ID":"e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0","Type":"ContainerStarted","Data":"21283897c3cb6eed7973a4aa7a52421e6e15bd217f32f70ce12aefb4fdd7a3ca"} Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.330332 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7825-account-create-phb7p" Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.362116 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g9pz\" (UniqueName: \"kubernetes.io/projected/835679c7-23a5-4347-b57c-eb0d600195b4-kube-api-access-8g9pz\") pod \"835679c7-23a5-4347-b57c-eb0d600195b4\" (UID: \"835679c7-23a5-4347-b57c-eb0d600195b4\") " Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.377120 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/835679c7-23a5-4347-b57c-eb0d600195b4-kube-api-access-8g9pz" (OuterVolumeSpecName: "kube-api-access-8g9pz") pod "835679c7-23a5-4347-b57c-eb0d600195b4" (UID: "835679c7-23a5-4347-b57c-eb0d600195b4"). InnerVolumeSpecName "kube-api-access-8g9pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.464060 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7555-account-create-rqrzd" Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.464305 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g9pz\" (UniqueName: \"kubernetes.io/projected/835679c7-23a5-4347-b57c-eb0d600195b4-kube-api-access-8g9pz\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.565816 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pxpms\" (UniqueName: \"kubernetes.io/projected/31ff6e52-d758-44a7-801c-7a0a9578157a-kube-api-access-pxpms\") pod \"31ff6e52-d758-44a7-801c-7a0a9578157a\" (UID: \"31ff6e52-d758-44a7-801c-7a0a9578157a\") " Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.570163 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31ff6e52-d758-44a7-801c-7a0a9578157a-kube-api-access-pxpms" (OuterVolumeSpecName: "kube-api-access-pxpms") pod "31ff6e52-d758-44a7-801c-7a0a9578157a" (UID: "31ff6e52-d758-44a7-801c-7a0a9578157a"). InnerVolumeSpecName "kube-api-access-pxpms". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:34 crc kubenswrapper[4923]: I1009 10:21:34.668243 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pxpms\" (UniqueName: \"kubernetes.io/projected/31ff6e52-d758-44a7-801c-7a0a9578157a-kube-api-access-pxpms\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.040611 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7555-account-create-rqrzd" event={"ID":"31ff6e52-d758-44a7-801c-7a0a9578157a","Type":"ContainerDied","Data":"52b81aa10441ad6d923d34ff88fc20f68c767dd59e258abf6826bebe8b9a2a65"} Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.040669 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52b81aa10441ad6d923d34ff88fc20f68c767dd59e258abf6826bebe8b9a2a65" Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.042895 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7555-account-create-rqrzd" Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.047386 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7825-account-create-phb7p" event={"ID":"835679c7-23a5-4347-b57c-eb0d600195b4","Type":"ContainerDied","Data":"8b7e59bcfb9fc2925e41ca54d354ba43502677c802e203c7328927e3d424a28f"} Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.047442 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b7e59bcfb9fc2925e41ca54d354ba43502677c802e203c7328927e3d424a28f" Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.047989 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7825-account-create-phb7p" Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.397714 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be11-account-create-9ms9b" Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.482400 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpfsf\" (UniqueName: \"kubernetes.io/projected/e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0-kube-api-access-jpfsf\") pod \"e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0\" (UID: \"e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0\") " Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.490173 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0-kube-api-access-jpfsf" (OuterVolumeSpecName: "kube-api-access-jpfsf") pod "e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0" (UID: "e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0"). InnerVolumeSpecName "kube-api-access-jpfsf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:35 crc kubenswrapper[4923]: I1009 10:21:35.584303 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpfsf\" (UniqueName: \"kubernetes.io/projected/e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0-kube-api-access-jpfsf\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:36 crc kubenswrapper[4923]: I1009 10:21:36.063226 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-be11-account-create-9ms9b" event={"ID":"e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0","Type":"ContainerDied","Data":"21283897c3cb6eed7973a4aa7a52421e6e15bd217f32f70ce12aefb4fdd7a3ca"} Oct 09 10:21:36 crc kubenswrapper[4923]: I1009 10:21:36.063303 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21283897c3cb6eed7973a4aa7a52421e6e15bd217f32f70ce12aefb4fdd7a3ca" Oct 09 10:21:36 crc kubenswrapper[4923]: I1009 10:21:36.063331 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-be11-account-create-9ms9b" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.124877 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jgs4j" podUID="4c323704-d81e-4c9b-a58d-b9943c2f55d7" containerName="ovn-controller" probeResult="failure" output=< Oct 09 10:21:37 crc kubenswrapper[4923]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 09 10:21:37 crc kubenswrapper[4923]: > Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.144723 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.416417 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-rmkl7"] Oct 09 10:21:37 crc kubenswrapper[4923]: E1009 10:21:37.416843 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0" containerName="mariadb-account-create" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.416868 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0" containerName="mariadb-account-create" Oct 09 10:21:37 crc kubenswrapper[4923]: E1009 10:21:37.416888 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31ff6e52-d758-44a7-801c-7a0a9578157a" containerName="mariadb-account-create" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.416896 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="31ff6e52-d758-44a7-801c-7a0a9578157a" containerName="mariadb-account-create" Oct 09 10:21:37 crc kubenswrapper[4923]: E1009 10:21:37.416909 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835679c7-23a5-4347-b57c-eb0d600195b4" containerName="mariadb-account-create" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.416917 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="835679c7-23a5-4347-b57c-eb0d600195b4" containerName="mariadb-account-create" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.417137 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="31ff6e52-d758-44a7-801c-7a0a9578157a" containerName="mariadb-account-create" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.417175 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0" containerName="mariadb-account-create" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.417205 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="835679c7-23a5-4347-b57c-eb0d600195b4" containerName="mariadb-account-create" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.417971 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.422345 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.422640 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mnjv6" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.424667 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rmkl7"] Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.517014 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-combined-ca-bundle\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.517063 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5cpfk\" (UniqueName: \"kubernetes.io/projected/f8febaa7-0d33-4ec9-90ce-07f415538366-kube-api-access-5cpfk\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.517129 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-config-data\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.517149 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-db-sync-config-data\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.619613 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-config-data\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.619709 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-db-sync-config-data\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.619911 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-combined-ca-bundle\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.619955 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5cpfk\" (UniqueName: \"kubernetes.io/projected/f8febaa7-0d33-4ec9-90ce-07f415538366-kube-api-access-5cpfk\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.627482 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-combined-ca-bundle\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.627538 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-config-data\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.628967 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-db-sync-config-data\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.637299 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5cpfk\" (UniqueName: \"kubernetes.io/projected/f8febaa7-0d33-4ec9-90ce-07f415538366-kube-api-access-5cpfk\") pod \"glance-db-sync-rmkl7\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:37 crc kubenswrapper[4923]: I1009 10:21:37.734864 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rmkl7" Oct 09 10:21:38 crc kubenswrapper[4923]: I1009 10:21:38.084326 4923 generic.go:334] "Generic (PLEG): container finished" podID="deeda18d-93f0-47c0-8cdd-358505802a0d" containerID="2049d58a742c0f3521d999b5b6d1678f1c3179e14b2ad167f60ebabd76e38a1f" exitCode=0 Oct 09 10:21:38 crc kubenswrapper[4923]: I1009 10:21:38.084527 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"deeda18d-93f0-47c0-8cdd-358505802a0d","Type":"ContainerDied","Data":"2049d58a742c0f3521d999b5b6d1678f1c3179e14b2ad167f60ebabd76e38a1f"} Oct 09 10:21:38 crc kubenswrapper[4923]: I1009 10:21:38.084848 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-rmkl7"] Oct 09 10:21:38 crc kubenswrapper[4923]: W1009 10:21:38.093810 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8febaa7_0d33_4ec9_90ce_07f415538366.slice/crio-c436cb7058bf9e2b3ce967a4afcdbd8fb207450093332fcbdad7449cec7ee5cc WatchSource:0}: Error finding container c436cb7058bf9e2b3ce967a4afcdbd8fb207450093332fcbdad7449cec7ee5cc: Status 404 returned error can't find the container with id c436cb7058bf9e2b3ce967a4afcdbd8fb207450093332fcbdad7449cec7ee5cc Oct 09 10:21:39 crc kubenswrapper[4923]: I1009 10:21:39.094195 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rmkl7" event={"ID":"f8febaa7-0d33-4ec9-90ce-07f415538366","Type":"ContainerStarted","Data":"c436cb7058bf9e2b3ce967a4afcdbd8fb207450093332fcbdad7449cec7ee5cc"} Oct 09 10:21:39 crc kubenswrapper[4923]: I1009 10:21:39.096555 4923 generic.go:334] "Generic (PLEG): container finished" podID="29358129-c9b5-4aca-bd00-0473171b3e96" containerID="863db0b0cc594cea2e1a5f5b6284f95afd669dcec8c8ece965c04e9aa52c6f22" exitCode=0 Oct 09 10:21:39 crc kubenswrapper[4923]: I1009 10:21:39.096654 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29358129-c9b5-4aca-bd00-0473171b3e96","Type":"ContainerDied","Data":"863db0b0cc594cea2e1a5f5b6284f95afd669dcec8c8ece965c04e9aa52c6f22"} Oct 09 10:21:39 crc kubenswrapper[4923]: I1009 10:21:39.100247 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"deeda18d-93f0-47c0-8cdd-358505802a0d","Type":"ContainerStarted","Data":"238b792fa9e2d3b834704439958b1cd2c20b1c82c8d08987fd20821005b85936"} Oct 09 10:21:39 crc kubenswrapper[4923]: I1009 10:21:39.100952 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 09 10:21:39 crc kubenswrapper[4923]: I1009 10:21:39.179044 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=52.741604023 podStartE2EDuration="1m2.179019148s" podCreationTimestamp="2025-10-09 10:20:37 +0000 UTC" firstStartedPulling="2025-10-09 10:20:53.763652805 +0000 UTC m=+939.831834561" lastFinishedPulling="2025-10-09 10:21:03.20106793 +0000 UTC m=+949.269249686" observedRunningTime="2025-10-09 10:21:39.167628262 +0000 UTC m=+985.235810018" watchObservedRunningTime="2025-10-09 10:21:39.179019148 +0000 UTC m=+985.247200904" Oct 09 10:21:40 crc kubenswrapper[4923]: I1009 10:21:40.113613 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29358129-c9b5-4aca-bd00-0473171b3e96","Type":"ContainerStarted","Data":"19cfaecaaad009dee38b2f7888556bee9b5a45ec2c3aa7f520e6d1a32e05c87b"} Oct 09 10:21:40 crc kubenswrapper[4923]: I1009 10:21:40.114250 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:21:40 crc kubenswrapper[4923]: I1009 10:21:40.153645 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=53.825481073 podStartE2EDuration="1m3.15362558s" podCreationTimestamp="2025-10-09 10:20:37 +0000 UTC" firstStartedPulling="2025-10-09 10:20:53.531586534 +0000 UTC m=+939.599768290" lastFinishedPulling="2025-10-09 10:21:02.859731041 +0000 UTC m=+948.927912797" observedRunningTime="2025-10-09 10:21:40.138434118 +0000 UTC m=+986.206615874" watchObservedRunningTime="2025-10-09 10:21:40.15362558 +0000 UTC m=+986.221807336" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.123934 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-jgs4j" podUID="4c323704-d81e-4c9b-a58d-b9943c2f55d7" containerName="ovn-controller" probeResult="failure" output=< Oct 09 10:21:42 crc kubenswrapper[4923]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 09 10:21:42 crc kubenswrapper[4923]: > Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.143658 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-m429c" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.360179 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-jgs4j-config-ctb92"] Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.361478 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.363735 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.381139 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgs4j-config-ctb92"] Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.526013 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-additional-scripts\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.526067 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-scripts\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.526144 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-log-ovn\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.526216 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94m4k\" (UniqueName: \"kubernetes.io/projected/1828e782-450f-493b-9a05-029d39b352a3-kube-api-access-94m4k\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.526236 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run-ovn\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.526263 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.628515 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.628626 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-additional-scripts\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.628673 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-scripts\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.628851 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-log-ovn\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.628925 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run-ovn\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.629085 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94m4k\" (UniqueName: \"kubernetes.io/projected/1828e782-450f-493b-9a05-029d39b352a3-kube-api-access-94m4k\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.629728 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.630456 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-additional-scripts\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.630845 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-log-ovn\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.630933 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run-ovn\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.634223 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-scripts\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.655202 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94m4k\" (UniqueName: \"kubernetes.io/projected/1828e782-450f-493b-9a05-029d39b352a3-kube-api-access-94m4k\") pod \"ovn-controller-jgs4j-config-ctb92\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:42 crc kubenswrapper[4923]: I1009 10:21:42.699422 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:43 crc kubenswrapper[4923]: I1009 10:21:43.193502 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-jgs4j-config-ctb92"] Oct 09 10:21:43 crc kubenswrapper[4923]: W1009 10:21:43.207217 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1828e782_450f_493b_9a05_029d39b352a3.slice/crio-c07a283101e0f8feb6536717b969b23ec1828a7abbd6ad99466fc527bd8e3e03 WatchSource:0}: Error finding container c07a283101e0f8feb6536717b969b23ec1828a7abbd6ad99466fc527bd8e3e03: Status 404 returned error can't find the container with id c07a283101e0f8feb6536717b969b23ec1828a7abbd6ad99466fc527bd8e3e03 Oct 09 10:21:44 crc kubenswrapper[4923]: I1009 10:21:44.164355 4923 generic.go:334] "Generic (PLEG): container finished" podID="1828e782-450f-493b-9a05-029d39b352a3" containerID="0720ad19d7d11b8e1b05aba153f1251872269ad2483f37c696bb36df0fa2f32b" exitCode=0 Oct 09 10:21:44 crc kubenswrapper[4923]: I1009 10:21:44.164437 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgs4j-config-ctb92" event={"ID":"1828e782-450f-493b-9a05-029d39b352a3","Type":"ContainerDied","Data":"0720ad19d7d11b8e1b05aba153f1251872269ad2483f37c696bb36df0fa2f32b"} Oct 09 10:21:44 crc kubenswrapper[4923]: I1009 10:21:44.165205 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgs4j-config-ctb92" event={"ID":"1828e782-450f-493b-9a05-029d39b352a3","Type":"ContainerStarted","Data":"c07a283101e0f8feb6536717b969b23ec1828a7abbd6ad99466fc527bd8e3e03"} Oct 09 10:21:47 crc kubenswrapper[4923]: I1009 10:21:47.119303 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-jgs4j" Oct 09 10:21:48 crc kubenswrapper[4923]: I1009 10:21:48.612457 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 09 10:21:48 crc kubenswrapper[4923]: I1009 10:21:48.987835 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-p77wh"] Oct 09 10:21:48 crc kubenswrapper[4923]: I1009 10:21:48.989546 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p77wh" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.004379 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-p77wh"] Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.006172 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="29358129-c9b5-4aca-bd00-0473171b3e96" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.095902 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-4545q"] Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.097432 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4545q" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.120180 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4545q"] Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.164071 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79jzm\" (UniqueName: \"kubernetes.io/projected/629aac6b-a0f1-4a58-b491-489a882cf65c-kube-api-access-79jzm\") pod \"barbican-db-create-p77wh\" (UID: \"629aac6b-a0f1-4a58-b491-489a882cf65c\") " pod="openstack/barbican-db-create-p77wh" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.266131 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79jzm\" (UniqueName: \"kubernetes.io/projected/629aac6b-a0f1-4a58-b491-489a882cf65c-kube-api-access-79jzm\") pod \"barbican-db-create-p77wh\" (UID: \"629aac6b-a0f1-4a58-b491-489a882cf65c\") " pod="openstack/barbican-db-create-p77wh" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.266238 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x96d6\" (UniqueName: \"kubernetes.io/projected/a1a8663c-24f3-4917-b301-4880439294df-kube-api-access-x96d6\") pod \"cinder-db-create-4545q\" (UID: \"a1a8663c-24f3-4917-b301-4880439294df\") " pod="openstack/cinder-db-create-4545q" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.290952 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-bwv6q"] Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.294837 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bwv6q" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.300142 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79jzm\" (UniqueName: \"kubernetes.io/projected/629aac6b-a0f1-4a58-b491-489a882cf65c-kube-api-access-79jzm\") pod \"barbican-db-create-p77wh\" (UID: \"629aac6b-a0f1-4a58-b491-489a882cf65c\") " pod="openstack/barbican-db-create-p77wh" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.305425 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bwv6q"] Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.368163 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x96d6\" (UniqueName: \"kubernetes.io/projected/a1a8663c-24f3-4917-b301-4880439294df-kube-api-access-x96d6\") pod \"cinder-db-create-4545q\" (UID: \"a1a8663c-24f3-4917-b301-4880439294df\") " pod="openstack/cinder-db-create-4545q" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.372745 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wbnxf"] Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.374382 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.380052 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.380328 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.380483 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.380544 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-58pgm" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.382346 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p77wh" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.394459 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wbnxf"] Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.425048 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x96d6\" (UniqueName: \"kubernetes.io/projected/a1a8663c-24f3-4917-b301-4880439294df-kube-api-access-x96d6\") pod \"cinder-db-create-4545q\" (UID: \"a1a8663c-24f3-4917-b301-4880439294df\") " pod="openstack/cinder-db-create-4545q" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.469926 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sd8x\" (UniqueName: \"kubernetes.io/projected/39c4d2f2-2bdd-417d-ab59-a99054774436-kube-api-access-9sd8x\") pod \"keystone-db-sync-wbnxf\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.469989 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-config-data\") pod \"keystone-db-sync-wbnxf\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.470024 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf6ck\" (UniqueName: \"kubernetes.io/projected/5e7d03c1-0cd8-41ca-aba6-880cf6fc2674-kube-api-access-cf6ck\") pod \"neutron-db-create-bwv6q\" (UID: \"5e7d03c1-0cd8-41ca-aba6-880cf6fc2674\") " pod="openstack/neutron-db-create-bwv6q" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.470086 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-combined-ca-bundle\") pod \"keystone-db-sync-wbnxf\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.571572 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sd8x\" (UniqueName: \"kubernetes.io/projected/39c4d2f2-2bdd-417d-ab59-a99054774436-kube-api-access-9sd8x\") pod \"keystone-db-sync-wbnxf\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.571633 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-config-data\") pod \"keystone-db-sync-wbnxf\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.571661 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf6ck\" (UniqueName: \"kubernetes.io/projected/5e7d03c1-0cd8-41ca-aba6-880cf6fc2674-kube-api-access-cf6ck\") pod \"neutron-db-create-bwv6q\" (UID: \"5e7d03c1-0cd8-41ca-aba6-880cf6fc2674\") " pod="openstack/neutron-db-create-bwv6q" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.571722 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-combined-ca-bundle\") pod \"keystone-db-sync-wbnxf\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.575001 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-combined-ca-bundle\") pod \"keystone-db-sync-wbnxf\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.576429 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-config-data\") pod \"keystone-db-sync-wbnxf\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.589916 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf6ck\" (UniqueName: \"kubernetes.io/projected/5e7d03c1-0cd8-41ca-aba6-880cf6fc2674-kube-api-access-cf6ck\") pod \"neutron-db-create-bwv6q\" (UID: \"5e7d03c1-0cd8-41ca-aba6-880cf6fc2674\") " pod="openstack/neutron-db-create-bwv6q" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.591457 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sd8x\" (UniqueName: \"kubernetes.io/projected/39c4d2f2-2bdd-417d-ab59-a99054774436-kube-api-access-9sd8x\") pod \"keystone-db-sync-wbnxf\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.664134 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bwv6q" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.696840 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:21:49 crc kubenswrapper[4923]: I1009 10:21:49.715282 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4545q" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.704200 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.816337 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-log-ovn\") pod \"1828e782-450f-493b-9a05-029d39b352a3\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.816398 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run-ovn\") pod \"1828e782-450f-493b-9a05-029d39b352a3\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.816478 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-scripts\") pod \"1828e782-450f-493b-9a05-029d39b352a3\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.816483 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1828e782-450f-493b-9a05-029d39b352a3" (UID: "1828e782-450f-493b-9a05-029d39b352a3"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.816537 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-additional-scripts\") pod \"1828e782-450f-493b-9a05-029d39b352a3\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.816578 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1828e782-450f-493b-9a05-029d39b352a3" (UID: "1828e782-450f-493b-9a05-029d39b352a3"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.816633 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run\") pod \"1828e782-450f-493b-9a05-029d39b352a3\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.816669 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94m4k\" (UniqueName: \"kubernetes.io/projected/1828e782-450f-493b-9a05-029d39b352a3-kube-api-access-94m4k\") pod \"1828e782-450f-493b-9a05-029d39b352a3\" (UID: \"1828e782-450f-493b-9a05-029d39b352a3\") " Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.816820 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run" (OuterVolumeSpecName: "var-run") pod "1828e782-450f-493b-9a05-029d39b352a3" (UID: "1828e782-450f-493b-9a05-029d39b352a3"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.817175 4923 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.817193 4923 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.817203 4923 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1828e782-450f-493b-9a05-029d39b352a3-var-run\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.817696 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1828e782-450f-493b-9a05-029d39b352a3" (UID: "1828e782-450f-493b-9a05-029d39b352a3"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.818031 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-scripts" (OuterVolumeSpecName: "scripts") pod "1828e782-450f-493b-9a05-029d39b352a3" (UID: "1828e782-450f-493b-9a05-029d39b352a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.825377 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1828e782-450f-493b-9a05-029d39b352a3-kube-api-access-94m4k" (OuterVolumeSpecName: "kube-api-access-94m4k") pod "1828e782-450f-493b-9a05-029d39b352a3" (UID: "1828e782-450f-493b-9a05-029d39b352a3"). InnerVolumeSpecName "kube-api-access-94m4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.919319 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.919360 4923 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1828e782-450f-493b-9a05-029d39b352a3-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:51 crc kubenswrapper[4923]: I1009 10:21:51.919374 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-94m4k\" (UniqueName: \"kubernetes.io/projected/1828e782-450f-493b-9a05-029d39b352a3-kube-api-access-94m4k\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.080990 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-bwv6q"] Oct 09 10:21:52 crc kubenswrapper[4923]: W1009 10:21:52.092558 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e7d03c1_0cd8_41ca_aba6_880cf6fc2674.slice/crio-4fb38e3bb574c01e588faaad79a55815560ec9b092c2bc588234139e6b250234 WatchSource:0}: Error finding container 4fb38e3bb574c01e588faaad79a55815560ec9b092c2bc588234139e6b250234: Status 404 returned error can't find the container with id 4fb38e3bb574c01e588faaad79a55815560ec9b092c2bc588234139e6b250234 Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.186228 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4545q"] Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.245742 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4545q" event={"ID":"a1a8663c-24f3-4917-b301-4880439294df","Type":"ContainerStarted","Data":"85f27f0e78eb05d964ba4104f7d832dfe6cabf6cf0ba95cc1cb1ce5c20295b79"} Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.247256 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-p77wh"] Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.249796 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-jgs4j-config-ctb92" Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.249842 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-jgs4j-config-ctb92" event={"ID":"1828e782-450f-493b-9a05-029d39b352a3","Type":"ContainerDied","Data":"c07a283101e0f8feb6536717b969b23ec1828a7abbd6ad99466fc527bd8e3e03"} Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.249894 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c07a283101e0f8feb6536717b969b23ec1828a7abbd6ad99466fc527bd8e3e03" Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.252043 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bwv6q" event={"ID":"5e7d03c1-0cd8-41ca-aba6-880cf6fc2674","Type":"ContainerStarted","Data":"4fb38e3bb574c01e588faaad79a55815560ec9b092c2bc588234139e6b250234"} Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.256564 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wbnxf"] Oct 09 10:21:52 crc kubenswrapper[4923]: W1009 10:21:52.261740 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod629aac6b_a0f1_4a58_b491_489a882cf65c.slice/crio-7901c2209094e666200ce56c23f4fa0a2bb43d815158c6e1b1f99a2a798c6987 WatchSource:0}: Error finding container 7901c2209094e666200ce56c23f4fa0a2bb43d815158c6e1b1f99a2a798c6987: Status 404 returned error can't find the container with id 7901c2209094e666200ce56c23f4fa0a2bb43d815158c6e1b1f99a2a798c6987 Oct 09 10:21:52 crc kubenswrapper[4923]: W1009 10:21:52.262456 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39c4d2f2_2bdd_417d_ab59_a99054774436.slice/crio-0abafd8655ce45da89d63d749ae6b88303e678441a9d8836c5d45ce00605642b WatchSource:0}: Error finding container 0abafd8655ce45da89d63d749ae6b88303e678441a9d8836c5d45ce00605642b: Status 404 returned error can't find the container with id 0abafd8655ce45da89d63d749ae6b88303e678441a9d8836c5d45ce00605642b Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.856923 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-jgs4j-config-ctb92"] Oct 09 10:21:52 crc kubenswrapper[4923]: I1009 10:21:52.874521 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-jgs4j-config-ctb92"] Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.289174 4923 generic.go:334] "Generic (PLEG): container finished" podID="629aac6b-a0f1-4a58-b491-489a882cf65c" containerID="cc98537e28beea7ef9b746322c1eeaf43f9ecbaf87fc70deefa5fba557733d0e" exitCode=0 Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.289309 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p77wh" event={"ID":"629aac6b-a0f1-4a58-b491-489a882cf65c","Type":"ContainerDied","Data":"cc98537e28beea7ef9b746322c1eeaf43f9ecbaf87fc70deefa5fba557733d0e"} Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.289357 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p77wh" event={"ID":"629aac6b-a0f1-4a58-b491-489a882cf65c","Type":"ContainerStarted","Data":"7901c2209094e666200ce56c23f4fa0a2bb43d815158c6e1b1f99a2a798c6987"} Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.290731 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wbnxf" event={"ID":"39c4d2f2-2bdd-417d-ab59-a99054774436","Type":"ContainerStarted","Data":"0abafd8655ce45da89d63d749ae6b88303e678441a9d8836c5d45ce00605642b"} Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.295229 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rmkl7" event={"ID":"f8febaa7-0d33-4ec9-90ce-07f415538366","Type":"ContainerStarted","Data":"d4e767d072551c1fdb7bb73cdd997ed2cdb0489605c1f682ea5f0080487fbe77"} Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.312558 4923 generic.go:334] "Generic (PLEG): container finished" podID="5e7d03c1-0cd8-41ca-aba6-880cf6fc2674" containerID="01b7f8b396fe8294d9d211391daeade41d79f6378e88876e0122b461e2b48c22" exitCode=0 Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.312694 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bwv6q" event={"ID":"5e7d03c1-0cd8-41ca-aba6-880cf6fc2674","Type":"ContainerDied","Data":"01b7f8b396fe8294d9d211391daeade41d79f6378e88876e0122b461e2b48c22"} Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.327690 4923 generic.go:334] "Generic (PLEG): container finished" podID="a1a8663c-24f3-4917-b301-4880439294df" containerID="4f9e9f9df6fd102dee0a5835a60d51311b8114f07ac5afb39e19889b26c73ab7" exitCode=0 Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.327824 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4545q" event={"ID":"a1a8663c-24f3-4917-b301-4880439294df","Type":"ContainerDied","Data":"4f9e9f9df6fd102dee0a5835a60d51311b8114f07ac5afb39e19889b26c73ab7"} Oct 09 10:21:53 crc kubenswrapper[4923]: I1009 10:21:53.350439 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-rmkl7" podStartSLOduration=2.77055591 podStartE2EDuration="16.350416667s" podCreationTimestamp="2025-10-09 10:21:37 +0000 UTC" firstStartedPulling="2025-10-09 10:21:38.096098106 +0000 UTC m=+984.164279862" lastFinishedPulling="2025-10-09 10:21:51.675958863 +0000 UTC m=+997.744140619" observedRunningTime="2025-10-09 10:21:53.334555778 +0000 UTC m=+999.402737524" watchObservedRunningTime="2025-10-09 10:21:53.350416667 +0000 UTC m=+999.418598423" Oct 09 10:21:54 crc kubenswrapper[4923]: I1009 10:21:54.619135 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1828e782-450f-493b-9a05-029d39b352a3" path="/var/lib/kubelet/pods/1828e782-450f-493b-9a05-029d39b352a3/volumes" Oct 09 10:21:54 crc kubenswrapper[4923]: I1009 10:21:54.715263 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bwv6q" Oct 09 10:21:54 crc kubenswrapper[4923]: I1009 10:21:54.809665 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4545q" Oct 09 10:21:54 crc kubenswrapper[4923]: I1009 10:21:54.817508 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p77wh" Oct 09 10:21:54 crc kubenswrapper[4923]: I1009 10:21:54.902267 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cf6ck\" (UniqueName: \"kubernetes.io/projected/5e7d03c1-0cd8-41ca-aba6-880cf6fc2674-kube-api-access-cf6ck\") pod \"5e7d03c1-0cd8-41ca-aba6-880cf6fc2674\" (UID: \"5e7d03c1-0cd8-41ca-aba6-880cf6fc2674\") " Oct 09 10:21:54 crc kubenswrapper[4923]: I1009 10:21:54.912055 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e7d03c1-0cd8-41ca-aba6-880cf6fc2674-kube-api-access-cf6ck" (OuterVolumeSpecName: "kube-api-access-cf6ck") pod "5e7d03c1-0cd8-41ca-aba6-880cf6fc2674" (UID: "5e7d03c1-0cd8-41ca-aba6-880cf6fc2674"). InnerVolumeSpecName "kube-api-access-cf6ck". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.004626 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79jzm\" (UniqueName: \"kubernetes.io/projected/629aac6b-a0f1-4a58-b491-489a882cf65c-kube-api-access-79jzm\") pod \"629aac6b-a0f1-4a58-b491-489a882cf65c\" (UID: \"629aac6b-a0f1-4a58-b491-489a882cf65c\") " Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.004682 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x96d6\" (UniqueName: \"kubernetes.io/projected/a1a8663c-24f3-4917-b301-4880439294df-kube-api-access-x96d6\") pod \"a1a8663c-24f3-4917-b301-4880439294df\" (UID: \"a1a8663c-24f3-4917-b301-4880439294df\") " Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.005038 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cf6ck\" (UniqueName: \"kubernetes.io/projected/5e7d03c1-0cd8-41ca-aba6-880cf6fc2674-kube-api-access-cf6ck\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.008925 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629aac6b-a0f1-4a58-b491-489a882cf65c-kube-api-access-79jzm" (OuterVolumeSpecName: "kube-api-access-79jzm") pod "629aac6b-a0f1-4a58-b491-489a882cf65c" (UID: "629aac6b-a0f1-4a58-b491-489a882cf65c"). InnerVolumeSpecName "kube-api-access-79jzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.009623 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a8663c-24f3-4917-b301-4880439294df-kube-api-access-x96d6" (OuterVolumeSpecName: "kube-api-access-x96d6") pod "a1a8663c-24f3-4917-b301-4880439294df" (UID: "a1a8663c-24f3-4917-b301-4880439294df"). InnerVolumeSpecName "kube-api-access-x96d6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.108562 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79jzm\" (UniqueName: \"kubernetes.io/projected/629aac6b-a0f1-4a58-b491-489a882cf65c-kube-api-access-79jzm\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.109196 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x96d6\" (UniqueName: \"kubernetes.io/projected/a1a8663c-24f3-4917-b301-4880439294df-kube-api-access-x96d6\") on node \"crc\" DevicePath \"\"" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.364391 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4545q" event={"ID":"a1a8663c-24f3-4917-b301-4880439294df","Type":"ContainerDied","Data":"85f27f0e78eb05d964ba4104f7d832dfe6cabf6cf0ba95cc1cb1ce5c20295b79"} Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.364452 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85f27f0e78eb05d964ba4104f7d832dfe6cabf6cf0ba95cc1cb1ce5c20295b79" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.364532 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4545q" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.368256 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-p77wh" event={"ID":"629aac6b-a0f1-4a58-b491-489a882cf65c","Type":"ContainerDied","Data":"7901c2209094e666200ce56c23f4fa0a2bb43d815158c6e1b1f99a2a798c6987"} Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.368290 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7901c2209094e666200ce56c23f4fa0a2bb43d815158c6e1b1f99a2a798c6987" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.368412 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-p77wh" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.371693 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-bwv6q" event={"ID":"5e7d03c1-0cd8-41ca-aba6-880cf6fc2674","Type":"ContainerDied","Data":"4fb38e3bb574c01e588faaad79a55815560ec9b092c2bc588234139e6b250234"} Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.371722 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4fb38e3bb574c01e588faaad79a55815560ec9b092c2bc588234139e6b250234" Oct 09 10:21:55 crc kubenswrapper[4923]: I1009 10:21:55.371822 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-bwv6q" Oct 09 10:21:59 crc kubenswrapper[4923]: I1009 10:21:59.005181 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:22:04 crc kubenswrapper[4923]: E1009 10:22:04.519187 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-keystone@sha256:1e5714637b6e1a24c2858fe6d9bbb3f00bc61d69ad74a657b1c23682bf4cb2b7" Oct 09 10:22:04 crc kubenswrapper[4923]: E1009 10:22:04.520064 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:keystone-db-sync,Image:quay.io/podified-antelope-centos9/openstack-keystone@sha256:1e5714637b6e1a24c2858fe6d9bbb3f00bc61d69ad74a657b1c23682bf4cb2b7,Command:[/bin/bash],Args:[-c keystone-manage db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/keystone/keystone.conf,SubPath:keystone.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9sd8x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42425,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42425,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-db-sync-wbnxf_openstack(39c4d2f2-2bdd-417d-ab59-a99054774436): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:22:04 crc kubenswrapper[4923]: E1009 10:22:04.521796 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/keystone-db-sync-wbnxf" podUID="39c4d2f2-2bdd-417d-ab59-a99054774436" Oct 09 10:22:05 crc kubenswrapper[4923]: E1009 10:22:05.530288 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"keystone-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-keystone@sha256:1e5714637b6e1a24c2858fe6d9bbb3f00bc61d69ad74a657b1c23682bf4cb2b7\\\"\"" pod="openstack/keystone-db-sync-wbnxf" podUID="39c4d2f2-2bdd-417d-ab59-a99054774436" Oct 09 10:22:06 crc kubenswrapper[4923]: I1009 10:22:06.518877 4923 generic.go:334] "Generic (PLEG): container finished" podID="f8febaa7-0d33-4ec9-90ce-07f415538366" containerID="d4e767d072551c1fdb7bb73cdd997ed2cdb0489605c1f682ea5f0080487fbe77" exitCode=0 Oct 09 10:22:06 crc kubenswrapper[4923]: I1009 10:22:06.519000 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rmkl7" event={"ID":"f8febaa7-0d33-4ec9-90ce-07f415538366","Type":"ContainerDied","Data":"d4e767d072551c1fdb7bb73cdd997ed2cdb0489605c1f682ea5f0080487fbe77"} Oct 09 10:22:07 crc kubenswrapper[4923]: I1009 10:22:07.978488 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rmkl7" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.157867 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-db-sync-config-data\") pod \"f8febaa7-0d33-4ec9-90ce-07f415538366\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.157948 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-combined-ca-bundle\") pod \"f8febaa7-0d33-4ec9-90ce-07f415538366\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.158004 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5cpfk\" (UniqueName: \"kubernetes.io/projected/f8febaa7-0d33-4ec9-90ce-07f415538366-kube-api-access-5cpfk\") pod \"f8febaa7-0d33-4ec9-90ce-07f415538366\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.158035 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-config-data\") pod \"f8febaa7-0d33-4ec9-90ce-07f415538366\" (UID: \"f8febaa7-0d33-4ec9-90ce-07f415538366\") " Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.166558 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "f8febaa7-0d33-4ec9-90ce-07f415538366" (UID: "f8febaa7-0d33-4ec9-90ce-07f415538366"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.170766 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8febaa7-0d33-4ec9-90ce-07f415538366-kube-api-access-5cpfk" (OuterVolumeSpecName: "kube-api-access-5cpfk") pod "f8febaa7-0d33-4ec9-90ce-07f415538366" (UID: "f8febaa7-0d33-4ec9-90ce-07f415538366"). InnerVolumeSpecName "kube-api-access-5cpfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.185437 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8febaa7-0d33-4ec9-90ce-07f415538366" (UID: "f8febaa7-0d33-4ec9-90ce-07f415538366"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.213817 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-config-data" (OuterVolumeSpecName: "config-data") pod "f8febaa7-0d33-4ec9-90ce-07f415538366" (UID: "f8febaa7-0d33-4ec9-90ce-07f415538366"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.260251 4923 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.260288 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.260298 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5cpfk\" (UniqueName: \"kubernetes.io/projected/f8febaa7-0d33-4ec9-90ce-07f415538366-kube-api-access-5cpfk\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.260309 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8febaa7-0d33-4ec9-90ce-07f415538366-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.539610 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-rmkl7" event={"ID":"f8febaa7-0d33-4ec9-90ce-07f415538366","Type":"ContainerDied","Data":"c436cb7058bf9e2b3ce967a4afcdbd8fb207450093332fcbdad7449cec7ee5cc"} Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.539999 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c436cb7058bf9e2b3ce967a4afcdbd8fb207450093332fcbdad7449cec7ee5cc" Oct 09 10:22:08 crc kubenswrapper[4923]: I1009 10:22:08.539697 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-rmkl7" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.003698 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bblwh"] Oct 09 10:22:09 crc kubenswrapper[4923]: E1009 10:22:09.004117 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1828e782-450f-493b-9a05-029d39b352a3" containerName="ovn-config" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.004137 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1828e782-450f-493b-9a05-029d39b352a3" containerName="ovn-config" Oct 09 10:22:09 crc kubenswrapper[4923]: E1009 10:22:09.004151 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629aac6b-a0f1-4a58-b491-489a882cf65c" containerName="mariadb-database-create" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.004158 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="629aac6b-a0f1-4a58-b491-489a882cf65c" containerName="mariadb-database-create" Oct 09 10:22:09 crc kubenswrapper[4923]: E1009 10:22:09.004179 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8febaa7-0d33-4ec9-90ce-07f415538366" containerName="glance-db-sync" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.004187 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8febaa7-0d33-4ec9-90ce-07f415538366" containerName="glance-db-sync" Oct 09 10:22:09 crc kubenswrapper[4923]: E1009 10:22:09.004203 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1a8663c-24f3-4917-b301-4880439294df" containerName="mariadb-database-create" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.004209 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1a8663c-24f3-4917-b301-4880439294df" containerName="mariadb-database-create" Oct 09 10:22:09 crc kubenswrapper[4923]: E1009 10:22:09.004219 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e7d03c1-0cd8-41ca-aba6-880cf6fc2674" containerName="mariadb-database-create" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.004224 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e7d03c1-0cd8-41ca-aba6-880cf6fc2674" containerName="mariadb-database-create" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.008572 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8febaa7-0d33-4ec9-90ce-07f415538366" containerName="glance-db-sync" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.008608 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1a8663c-24f3-4917-b301-4880439294df" containerName="mariadb-database-create" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.008623 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="1828e782-450f-493b-9a05-029d39b352a3" containerName="ovn-config" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.008636 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="629aac6b-a0f1-4a58-b491-489a882cf65c" containerName="mariadb-database-create" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.008653 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e7d03c1-0cd8-41ca-aba6-880cf6fc2674" containerName="mariadb-database-create" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.009697 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.032733 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bblwh"] Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.098532 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-sb\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.098596 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-nb\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.098852 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh7c5\" (UniqueName: \"kubernetes.io/projected/255708b2-d7b2-4f19-9ae2-5cf29d005025-kube-api-access-rh7c5\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.099116 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-dns-svc\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.099258 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-config\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.104342 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-51be-account-create-dswv4"] Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.105774 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-51be-account-create-dswv4" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.109622 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.122312 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-51be-account-create-dswv4"] Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.201193 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-config\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.201292 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-sb\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.201323 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-nb\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.201383 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh7c5\" (UniqueName: \"kubernetes.io/projected/255708b2-d7b2-4f19-9ae2-5cf29d005025-kube-api-access-rh7c5\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.201431 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bnhh\" (UniqueName: \"kubernetes.io/projected/d3151f13-c711-4e82-b24b-53be5ecf3e66-kube-api-access-9bnhh\") pod \"barbican-51be-account-create-dswv4\" (UID: \"d3151f13-c711-4e82-b24b-53be5ecf3e66\") " pod="openstack/barbican-51be-account-create-dswv4" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.201478 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-dns-svc\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.202653 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-config\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.202665 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-sb\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.202801 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-nb\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.202973 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-dns-svc\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.227507 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh7c5\" (UniqueName: \"kubernetes.io/projected/255708b2-d7b2-4f19-9ae2-5cf29d005025-kube-api-access-rh7c5\") pod \"dnsmasq-dns-74b7749bc7-bblwh\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.303516 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bnhh\" (UniqueName: \"kubernetes.io/projected/d3151f13-c711-4e82-b24b-53be5ecf3e66-kube-api-access-9bnhh\") pod \"barbican-51be-account-create-dswv4\" (UID: \"d3151f13-c711-4e82-b24b-53be5ecf3e66\") " pod="openstack/barbican-51be-account-create-dswv4" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.321205 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-3a12-account-create-wsttd"] Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.322808 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3a12-account-create-wsttd" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.327302 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.333119 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bnhh\" (UniqueName: \"kubernetes.io/projected/d3151f13-c711-4e82-b24b-53be5ecf3e66-kube-api-access-9bnhh\") pod \"barbican-51be-account-create-dswv4\" (UID: \"d3151f13-c711-4e82-b24b-53be5ecf3e66\") " pod="openstack/barbican-51be-account-create-dswv4" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.341179 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3a12-account-create-wsttd"] Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.350189 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.406428 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6m967\" (UniqueName: \"kubernetes.io/projected/3bede9ca-eff0-422f-bc53-b156346d993f-kube-api-access-6m967\") pod \"cinder-3a12-account-create-wsttd\" (UID: \"3bede9ca-eff0-422f-bc53-b156346d993f\") " pod="openstack/cinder-3a12-account-create-wsttd" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.427232 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-51be-account-create-dswv4" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.509158 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6m967\" (UniqueName: \"kubernetes.io/projected/3bede9ca-eff0-422f-bc53-b156346d993f-kube-api-access-6m967\") pod \"cinder-3a12-account-create-wsttd\" (UID: \"3bede9ca-eff0-422f-bc53-b156346d993f\") " pod="openstack/cinder-3a12-account-create-wsttd" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.523702 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-55d5-account-create-wsrr5"] Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.525123 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55d5-account-create-wsrr5" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.528474 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.536445 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55d5-account-create-wsrr5"] Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.544182 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6m967\" (UniqueName: \"kubernetes.io/projected/3bede9ca-eff0-422f-bc53-b156346d993f-kube-api-access-6m967\") pod \"cinder-3a12-account-create-wsttd\" (UID: \"3bede9ca-eff0-422f-bc53-b156346d993f\") " pod="openstack/cinder-3a12-account-create-wsttd" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.610334 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b876v\" (UniqueName: \"kubernetes.io/projected/ee194426-1c9c-49f6-b112-8bf243b2d8ca-kube-api-access-b876v\") pod \"neutron-55d5-account-create-wsrr5\" (UID: \"ee194426-1c9c-49f6-b112-8bf243b2d8ca\") " pod="openstack/neutron-55d5-account-create-wsrr5" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.714471 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b876v\" (UniqueName: \"kubernetes.io/projected/ee194426-1c9c-49f6-b112-8bf243b2d8ca-kube-api-access-b876v\") pod \"neutron-55d5-account-create-wsrr5\" (UID: \"ee194426-1c9c-49f6-b112-8bf243b2d8ca\") " pod="openstack/neutron-55d5-account-create-wsrr5" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.759677 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b876v\" (UniqueName: \"kubernetes.io/projected/ee194426-1c9c-49f6-b112-8bf243b2d8ca-kube-api-access-b876v\") pod \"neutron-55d5-account-create-wsrr5\" (UID: \"ee194426-1c9c-49f6-b112-8bf243b2d8ca\") " pod="openstack/neutron-55d5-account-create-wsrr5" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.800013 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3a12-account-create-wsttd" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.800389 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-51be-account-create-dswv4"] Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.854171 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55d5-account-create-wsrr5" Oct 09 10:22:09 crc kubenswrapper[4923]: I1009 10:22:09.914505 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bblwh"] Oct 09 10:22:10 crc kubenswrapper[4923]: I1009 10:22:10.252066 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-3a12-account-create-wsttd"] Oct 09 10:22:10 crc kubenswrapper[4923]: W1009 10:22:10.267073 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bede9ca_eff0_422f_bc53_b156346d993f.slice/crio-f0c6a45c9910e32e8c6f2681bc2b3bd495303df2c9639f81cf82270fb5ce1387 WatchSource:0}: Error finding container f0c6a45c9910e32e8c6f2681bc2b3bd495303df2c9639f81cf82270fb5ce1387: Status 404 returned error can't find the container with id f0c6a45c9910e32e8c6f2681bc2b3bd495303df2c9639f81cf82270fb5ce1387 Oct 09 10:22:10 crc kubenswrapper[4923]: I1009 10:22:10.564071 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-55d5-account-create-wsrr5"] Oct 09 10:22:10 crc kubenswrapper[4923]: I1009 10:22:10.576073 4923 generic.go:334] "Generic (PLEG): container finished" podID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerID="73e71e1a63690f76fc38a9e0bcbc28a007f2b6c091a88d0b097f12ced24139c7" exitCode=0 Oct 09 10:22:10 crc kubenswrapper[4923]: I1009 10:22:10.576105 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" event={"ID":"255708b2-d7b2-4f19-9ae2-5cf29d005025","Type":"ContainerDied","Data":"73e71e1a63690f76fc38a9e0bcbc28a007f2b6c091a88d0b097f12ced24139c7"} Oct 09 10:22:10 crc kubenswrapper[4923]: I1009 10:22:10.576144 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" event={"ID":"255708b2-d7b2-4f19-9ae2-5cf29d005025","Type":"ContainerStarted","Data":"6d1b53d8fb47f058a8168e207ec84a86ef40b240a3c36bd0d6945bfd7d17bc8d"} Oct 09 10:22:10 crc kubenswrapper[4923]: I1009 10:22:10.580367 4923 generic.go:334] "Generic (PLEG): container finished" podID="d3151f13-c711-4e82-b24b-53be5ecf3e66" containerID="dfc10e4bff23b6038dbd423ffc673adfbd68742eb983ad8f2ee294ce53c39893" exitCode=0 Oct 09 10:22:10 crc kubenswrapper[4923]: I1009 10:22:10.580446 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-51be-account-create-dswv4" event={"ID":"d3151f13-c711-4e82-b24b-53be5ecf3e66","Type":"ContainerDied","Data":"dfc10e4bff23b6038dbd423ffc673adfbd68742eb983ad8f2ee294ce53c39893"} Oct 09 10:22:10 crc kubenswrapper[4923]: I1009 10:22:10.580479 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-51be-account-create-dswv4" event={"ID":"d3151f13-c711-4e82-b24b-53be5ecf3e66","Type":"ContainerStarted","Data":"91a9bcf22c7ba5af70a880062dc56ae04d41bb086f0fb7562a27630bf0aa2516"} Oct 09 10:22:10 crc kubenswrapper[4923]: I1009 10:22:10.587118 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3a12-account-create-wsttd" event={"ID":"3bede9ca-eff0-422f-bc53-b156346d993f","Type":"ContainerStarted","Data":"f0c6a45c9910e32e8c6f2681bc2b3bd495303df2c9639f81cf82270fb5ce1387"} Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.616976 4923 generic.go:334] "Generic (PLEG): container finished" podID="ee194426-1c9c-49f6-b112-8bf243b2d8ca" containerID="52dfbd33a4b04e87362c2cc041044c5e407650b00fe639046e1159b961498545" exitCode=0 Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.617594 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55d5-account-create-wsrr5" event={"ID":"ee194426-1c9c-49f6-b112-8bf243b2d8ca","Type":"ContainerDied","Data":"52dfbd33a4b04e87362c2cc041044c5e407650b00fe639046e1159b961498545"} Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.617644 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55d5-account-create-wsrr5" event={"ID":"ee194426-1c9c-49f6-b112-8bf243b2d8ca","Type":"ContainerStarted","Data":"7b143506dde6bd9ba2ee306f8a1768222953cf83e6ed07416e8fc788272d223e"} Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.635342 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" event={"ID":"255708b2-d7b2-4f19-9ae2-5cf29d005025","Type":"ContainerStarted","Data":"bc03f56b265729775f1504bcc57cd7ad1d38ee3f4d52e0a69914696894dd5d01"} Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.635403 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.639484 4923 generic.go:334] "Generic (PLEG): container finished" podID="3bede9ca-eff0-422f-bc53-b156346d993f" containerID="61c0e03f87d6e78e0d9732565997c372418a8a985104a15d889346157ca27e70" exitCode=0 Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.639713 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3a12-account-create-wsttd" event={"ID":"3bede9ca-eff0-422f-bc53-b156346d993f","Type":"ContainerDied","Data":"61c0e03f87d6e78e0d9732565997c372418a8a985104a15d889346157ca27e70"} Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.707453 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" podStartSLOduration=3.707424282 podStartE2EDuration="3.707424282s" podCreationTimestamp="2025-10-09 10:22:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:22:11.69253064 +0000 UTC m=+1017.760712396" watchObservedRunningTime="2025-10-09 10:22:11.707424282 +0000 UTC m=+1017.775606038" Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.939790 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-51be-account-create-dswv4" Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.974595 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bnhh\" (UniqueName: \"kubernetes.io/projected/d3151f13-c711-4e82-b24b-53be5ecf3e66-kube-api-access-9bnhh\") pod \"d3151f13-c711-4e82-b24b-53be5ecf3e66\" (UID: \"d3151f13-c711-4e82-b24b-53be5ecf3e66\") " Oct 09 10:22:11 crc kubenswrapper[4923]: I1009 10:22:11.981850 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3151f13-c711-4e82-b24b-53be5ecf3e66-kube-api-access-9bnhh" (OuterVolumeSpecName: "kube-api-access-9bnhh") pod "d3151f13-c711-4e82-b24b-53be5ecf3e66" (UID: "d3151f13-c711-4e82-b24b-53be5ecf3e66"). InnerVolumeSpecName "kube-api-access-9bnhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:12 crc kubenswrapper[4923]: I1009 10:22:12.077387 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bnhh\" (UniqueName: \"kubernetes.io/projected/d3151f13-c711-4e82-b24b-53be5ecf3e66-kube-api-access-9bnhh\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:12 crc kubenswrapper[4923]: I1009 10:22:12.650961 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-51be-account-create-dswv4" event={"ID":"d3151f13-c711-4e82-b24b-53be5ecf3e66","Type":"ContainerDied","Data":"91a9bcf22c7ba5af70a880062dc56ae04d41bb086f0fb7562a27630bf0aa2516"} Oct 09 10:22:12 crc kubenswrapper[4923]: I1009 10:22:12.651036 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91a9bcf22c7ba5af70a880062dc56ae04d41bb086f0fb7562a27630bf0aa2516" Oct 09 10:22:12 crc kubenswrapper[4923]: I1009 10:22:12.653496 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-51be-account-create-dswv4" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.070027 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3a12-account-create-wsttd" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.075806 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55d5-account-create-wsrr5" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.096089 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6m967\" (UniqueName: \"kubernetes.io/projected/3bede9ca-eff0-422f-bc53-b156346d993f-kube-api-access-6m967\") pod \"3bede9ca-eff0-422f-bc53-b156346d993f\" (UID: \"3bede9ca-eff0-422f-bc53-b156346d993f\") " Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.096453 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b876v\" (UniqueName: \"kubernetes.io/projected/ee194426-1c9c-49f6-b112-8bf243b2d8ca-kube-api-access-b876v\") pod \"ee194426-1c9c-49f6-b112-8bf243b2d8ca\" (UID: \"ee194426-1c9c-49f6-b112-8bf243b2d8ca\") " Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.103960 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee194426-1c9c-49f6-b112-8bf243b2d8ca-kube-api-access-b876v" (OuterVolumeSpecName: "kube-api-access-b876v") pod "ee194426-1c9c-49f6-b112-8bf243b2d8ca" (UID: "ee194426-1c9c-49f6-b112-8bf243b2d8ca"). InnerVolumeSpecName "kube-api-access-b876v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.106871 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bede9ca-eff0-422f-bc53-b156346d993f-kube-api-access-6m967" (OuterVolumeSpecName: "kube-api-access-6m967") pod "3bede9ca-eff0-422f-bc53-b156346d993f" (UID: "3bede9ca-eff0-422f-bc53-b156346d993f"). InnerVolumeSpecName "kube-api-access-6m967". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.198830 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b876v\" (UniqueName: \"kubernetes.io/projected/ee194426-1c9c-49f6-b112-8bf243b2d8ca-kube-api-access-b876v\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.198891 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6m967\" (UniqueName: \"kubernetes.io/projected/3bede9ca-eff0-422f-bc53-b156346d993f-kube-api-access-6m967\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.666041 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-3a12-account-create-wsttd" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.666043 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-3a12-account-create-wsttd" event={"ID":"3bede9ca-eff0-422f-bc53-b156346d993f","Type":"ContainerDied","Data":"f0c6a45c9910e32e8c6f2681bc2b3bd495303df2c9639f81cf82270fb5ce1387"} Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.667321 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0c6a45c9910e32e8c6f2681bc2b3bd495303df2c9639f81cf82270fb5ce1387" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.668242 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-55d5-account-create-wsrr5" event={"ID":"ee194426-1c9c-49f6-b112-8bf243b2d8ca","Type":"ContainerDied","Data":"7b143506dde6bd9ba2ee306f8a1768222953cf83e6ed07416e8fc788272d223e"} Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.668279 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b143506dde6bd9ba2ee306f8a1768222953cf83e6ed07416e8fc788272d223e" Oct 09 10:22:13 crc kubenswrapper[4923]: I1009 10:22:13.668337 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-55d5-account-create-wsrr5" Oct 09 10:22:17 crc kubenswrapper[4923]: I1009 10:22:17.701894 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wbnxf" event={"ID":"39c4d2f2-2bdd-417d-ab59-a99054774436","Type":"ContainerStarted","Data":"d8f0aa1852932b3fb1536d7b44c02302c13c8c4791f9719aca6fd74693698c44"} Oct 09 10:22:17 crc kubenswrapper[4923]: I1009 10:22:17.724364 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wbnxf" podStartSLOduration=4.490553068 podStartE2EDuration="28.724345132s" podCreationTimestamp="2025-10-09 10:21:49 +0000 UTC" firstStartedPulling="2025-10-09 10:21:52.268996887 +0000 UTC m=+998.337178643" lastFinishedPulling="2025-10-09 10:22:16.502788951 +0000 UTC m=+1022.570970707" observedRunningTime="2025-10-09 10:22:17.717278647 +0000 UTC m=+1023.785460403" watchObservedRunningTime="2025-10-09 10:22:17.724345132 +0000 UTC m=+1023.792526878" Oct 09 10:22:19 crc kubenswrapper[4923]: I1009 10:22:19.352018 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:19 crc kubenswrapper[4923]: I1009 10:22:19.405210 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-wtm94"] Oct 09 10:22:19 crc kubenswrapper[4923]: I1009 10:22:19.405506 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" podUID="1eb424ba-b604-47d9-88f0-c268ccdd5aa3" containerName="dnsmasq-dns" containerID="cri-o://53e977ac3d7dba4288d2e45bacf549b66fee42a44884f8b0403e44dfc8108f39" gracePeriod=10 Oct 09 10:22:19 crc kubenswrapper[4923]: I1009 10:22:19.718083 4923 generic.go:334] "Generic (PLEG): container finished" podID="1eb424ba-b604-47d9-88f0-c268ccdd5aa3" containerID="53e977ac3d7dba4288d2e45bacf549b66fee42a44884f8b0403e44dfc8108f39" exitCode=0 Oct 09 10:22:19 crc kubenswrapper[4923]: I1009 10:22:19.718135 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" event={"ID":"1eb424ba-b604-47d9-88f0-c268ccdd5aa3","Type":"ContainerDied","Data":"53e977ac3d7dba4288d2e45bacf549b66fee42a44884f8b0403e44dfc8108f39"} Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.095070 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.138625 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-nb\") pod \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.138783 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7c5h\" (UniqueName: \"kubernetes.io/projected/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-kube-api-access-t7c5h\") pod \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.138879 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-dns-svc\") pod \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.139043 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-sb\") pod \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.139183 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-config\") pod \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\" (UID: \"1eb424ba-b604-47d9-88f0-c268ccdd5aa3\") " Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.149286 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-kube-api-access-t7c5h" (OuterVolumeSpecName: "kube-api-access-t7c5h") pod "1eb424ba-b604-47d9-88f0-c268ccdd5aa3" (UID: "1eb424ba-b604-47d9-88f0-c268ccdd5aa3"). InnerVolumeSpecName "kube-api-access-t7c5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.190370 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1eb424ba-b604-47d9-88f0-c268ccdd5aa3" (UID: "1eb424ba-b604-47d9-88f0-c268ccdd5aa3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.198052 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-config" (OuterVolumeSpecName: "config") pod "1eb424ba-b604-47d9-88f0-c268ccdd5aa3" (UID: "1eb424ba-b604-47d9-88f0-c268ccdd5aa3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.203880 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1eb424ba-b604-47d9-88f0-c268ccdd5aa3" (UID: "1eb424ba-b604-47d9-88f0-c268ccdd5aa3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.205500 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1eb424ba-b604-47d9-88f0-c268ccdd5aa3" (UID: "1eb424ba-b604-47d9-88f0-c268ccdd5aa3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.242619 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.242661 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.242672 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.242683 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7c5h\" (UniqueName: \"kubernetes.io/projected/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-kube-api-access-t7c5h\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.242694 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1eb424ba-b604-47d9-88f0-c268ccdd5aa3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.728516 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" event={"ID":"1eb424ba-b604-47d9-88f0-c268ccdd5aa3","Type":"ContainerDied","Data":"5a8d2a2c0846a6f4d3fcd38566c524b52550797ba733600f61cbc9e04807ac30"} Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.728579 4923 scope.go:117] "RemoveContainer" containerID="53e977ac3d7dba4288d2e45bacf549b66fee42a44884f8b0403e44dfc8108f39" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.728656 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dc9d58d7-wtm94" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.729817 4923 generic.go:334] "Generic (PLEG): container finished" podID="39c4d2f2-2bdd-417d-ab59-a99054774436" containerID="d8f0aa1852932b3fb1536d7b44c02302c13c8c4791f9719aca6fd74693698c44" exitCode=0 Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.729855 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wbnxf" event={"ID":"39c4d2f2-2bdd-417d-ab59-a99054774436","Type":"ContainerDied","Data":"d8f0aa1852932b3fb1536d7b44c02302c13c8c4791f9719aca6fd74693698c44"} Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.766235 4923 scope.go:117] "RemoveContainer" containerID="03e098255c35400936fd2b183de45fe3394363fcff61b95395f96f6875f29384" Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.779234 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-wtm94"] Oct 09 10:22:20 crc kubenswrapper[4923]: I1009 10:22:20.786879 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dc9d58d7-wtm94"] Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.062837 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.173720 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-config-data\") pod \"39c4d2f2-2bdd-417d-ab59-a99054774436\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.173944 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sd8x\" (UniqueName: \"kubernetes.io/projected/39c4d2f2-2bdd-417d-ab59-a99054774436-kube-api-access-9sd8x\") pod \"39c4d2f2-2bdd-417d-ab59-a99054774436\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.173978 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-combined-ca-bundle\") pod \"39c4d2f2-2bdd-417d-ab59-a99054774436\" (UID: \"39c4d2f2-2bdd-417d-ab59-a99054774436\") " Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.181180 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39c4d2f2-2bdd-417d-ab59-a99054774436-kube-api-access-9sd8x" (OuterVolumeSpecName: "kube-api-access-9sd8x") pod "39c4d2f2-2bdd-417d-ab59-a99054774436" (UID: "39c4d2f2-2bdd-417d-ab59-a99054774436"). InnerVolumeSpecName "kube-api-access-9sd8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.202698 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "39c4d2f2-2bdd-417d-ab59-a99054774436" (UID: "39c4d2f2-2bdd-417d-ab59-a99054774436"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.224175 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-config-data" (OuterVolumeSpecName: "config-data") pod "39c4d2f2-2bdd-417d-ab59-a99054774436" (UID: "39c4d2f2-2bdd-417d-ab59-a99054774436"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.276144 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sd8x\" (UniqueName: \"kubernetes.io/projected/39c4d2f2-2bdd-417d-ab59-a99054774436-kube-api-access-9sd8x\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.276182 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.276192 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/39c4d2f2-2bdd-417d-ab59-a99054774436-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.614628 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eb424ba-b604-47d9-88f0-c268ccdd5aa3" path="/var/lib/kubelet/pods/1eb424ba-b604-47d9-88f0-c268ccdd5aa3/volumes" Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.747149 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wbnxf" event={"ID":"39c4d2f2-2bdd-417d-ab59-a99054774436","Type":"ContainerDied","Data":"0abafd8655ce45da89d63d749ae6b88303e678441a9d8836c5d45ce00605642b"} Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.747207 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0abafd8655ce45da89d63d749ae6b88303e678441a9d8836c5d45ce00605642b" Oct 09 10:22:22 crc kubenswrapper[4923]: I1009 10:22:22.747263 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wbnxf" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.020699 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-nxfq4"] Oct 09 10:22:23 crc kubenswrapper[4923]: E1009 10:22:23.021210 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bede9ca-eff0-422f-bc53-b156346d993f" containerName="mariadb-account-create" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021238 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bede9ca-eff0-422f-bc53-b156346d993f" containerName="mariadb-account-create" Oct 09 10:22:23 crc kubenswrapper[4923]: E1009 10:22:23.021251 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39c4d2f2-2bdd-417d-ab59-a99054774436" containerName="keystone-db-sync" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021258 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="39c4d2f2-2bdd-417d-ab59-a99054774436" containerName="keystone-db-sync" Oct 09 10:22:23 crc kubenswrapper[4923]: E1009 10:22:23.021277 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3151f13-c711-4e82-b24b-53be5ecf3e66" containerName="mariadb-account-create" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021285 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3151f13-c711-4e82-b24b-53be5ecf3e66" containerName="mariadb-account-create" Oct 09 10:22:23 crc kubenswrapper[4923]: E1009 10:22:23.021294 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb424ba-b604-47d9-88f0-c268ccdd5aa3" containerName="dnsmasq-dns" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021300 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb424ba-b604-47d9-88f0-c268ccdd5aa3" containerName="dnsmasq-dns" Oct 09 10:22:23 crc kubenswrapper[4923]: E1009 10:22:23.021311 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb424ba-b604-47d9-88f0-c268ccdd5aa3" containerName="init" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021317 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb424ba-b604-47d9-88f0-c268ccdd5aa3" containerName="init" Oct 09 10:22:23 crc kubenswrapper[4923]: E1009 10:22:23.021329 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee194426-1c9c-49f6-b112-8bf243b2d8ca" containerName="mariadb-account-create" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021334 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee194426-1c9c-49f6-b112-8bf243b2d8ca" containerName="mariadb-account-create" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021491 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee194426-1c9c-49f6-b112-8bf243b2d8ca" containerName="mariadb-account-create" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021521 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3151f13-c711-4e82-b24b-53be5ecf3e66" containerName="mariadb-account-create" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021531 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb424ba-b604-47d9-88f0-c268ccdd5aa3" containerName="dnsmasq-dns" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021545 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="39c4d2f2-2bdd-417d-ab59-a99054774436" containerName="keystone-db-sync" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.021554 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bede9ca-eff0-422f-bc53-b156346d993f" containerName="mariadb-account-create" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.022957 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.045022 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-nxfq4"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.068098 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5fbrp"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.069571 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.081242 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.081585 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.081889 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-58pgm" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.082059 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.091212 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-sb\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.091592 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-nb\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.091632 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-config\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.091660 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgbxw\" (UniqueName: \"kubernetes.io/projected/c7fc5632-a21c-436c-8a74-7c99de29915d-kube-api-access-jgbxw\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.091784 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-dns-svc\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.091902 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5fbrp"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197015 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4mwc\" (UniqueName: \"kubernetes.io/projected/9fb0fa45-d830-4b81-9383-39b8e622fe0c-kube-api-access-f4mwc\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197103 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-sb\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197156 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-scripts\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197179 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-credential-keys\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197209 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-config-data\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197234 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-nb\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197272 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-config\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197296 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgbxw\" (UniqueName: \"kubernetes.io/projected/c7fc5632-a21c-436c-8a74-7c99de29915d-kube-api-access-jgbxw\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197342 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-fernet-keys\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197396 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-combined-ca-bundle\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.197457 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-dns-svc\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.198595 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-dns-svc\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.199006 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-sb\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.199255 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-nb\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.199837 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-config\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.250170 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-776bc5df9c-2wltl"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.252516 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.256909 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-q4cgz" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.257381 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.260223 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.260692 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.263046 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgbxw\" (UniqueName: \"kubernetes.io/projected/c7fc5632-a21c-436c-8a74-7c99de29915d-kube-api-access-jgbxw\") pod \"dnsmasq-dns-67bcfd764f-nxfq4\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.281221 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-5ll8s"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.282413 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.286690 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.286855 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.287427 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kzbkw" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.299285 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-fernet-keys\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.299355 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-combined-ca-bundle\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.299417 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wszh4\" (UniqueName: \"kubernetes.io/projected/75eb01a3-916a-4825-a8b4-1de9970e4e6e-kube-api-access-wszh4\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.299447 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75eb01a3-916a-4825-a8b4-1de9970e4e6e-logs\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.299534 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4mwc\" (UniqueName: \"kubernetes.io/projected/9fb0fa45-d830-4b81-9383-39b8e622fe0c-kube-api-access-f4mwc\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.299577 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-scripts\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.303991 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-credential-keys\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.304091 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-config-data\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.304141 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75eb01a3-916a-4825-a8b4-1de9970e4e6e-horizon-secret-key\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.304208 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-config-data\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.304241 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-scripts\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.316820 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-776bc5df9c-2wltl"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.320450 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-scripts\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.326233 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-config-data\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.330111 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-combined-ca-bundle\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.333503 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-credential-keys\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.339887 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-fernet-keys\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.344812 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4mwc\" (UniqueName: \"kubernetes.io/projected/9fb0fa45-d830-4b81-9383-39b8e622fe0c-kube-api-access-f4mwc\") pod \"keystone-bootstrap-5fbrp\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.347059 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.407897 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5ll8s"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409490 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75eb01a3-916a-4825-a8b4-1de9970e4e6e-logs\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409604 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-db-sync-config-data\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409675 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-combined-ca-bundle\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409745 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7rgc\" (UniqueName: \"kubernetes.io/projected/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-kube-api-access-r7rgc\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409800 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-etc-machine-id\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409825 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-config-data\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409863 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75eb01a3-916a-4825-a8b4-1de9970e4e6e-horizon-secret-key\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409907 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-config-data\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409934 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-scripts\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.409976 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-scripts\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.410068 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wszh4\" (UniqueName: \"kubernetes.io/projected/75eb01a3-916a-4825-a8b4-1de9970e4e6e-kube-api-access-wszh4\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.416308 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75eb01a3-916a-4825-a8b4-1de9970e4e6e-logs\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.422479 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-scripts\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.429834 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-config-data\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.430267 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.445202 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wszh4\" (UniqueName: \"kubernetes.io/projected/75eb01a3-916a-4825-a8b4-1de9970e4e6e-kube-api-access-wszh4\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.459261 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75eb01a3-916a-4825-a8b4-1de9970e4e6e-horizon-secret-key\") pod \"horizon-776bc5df9c-2wltl\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.520482 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-db-sync-config-data\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.520648 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-combined-ca-bundle\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.520739 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7rgc\" (UniqueName: \"kubernetes.io/projected/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-kube-api-access-r7rgc\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.520788 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-etc-machine-id\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.520810 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-config-data\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.521024 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-scripts\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.531455 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-etc-machine-id\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.547724 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-combined-ca-bundle\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.547823 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-5jlhb"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.547839 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-db-sync-config-data\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.548058 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-config-data\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.549243 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.553341 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.554621 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-c5hbk" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.561716 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.568193 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-scripts\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.579434 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7rgc\" (UniqueName: \"kubernetes.io/projected/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-kube-api-access-r7rgc\") pod \"cinder-db-sync-5ll8s\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.580290 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5jlhb"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.607207 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.609404 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.614639 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.614873 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.630615 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.635054 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-config\") pod \"neutron-db-sync-5jlhb\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.635778 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-combined-ca-bundle\") pod \"neutron-db-sync-5jlhb\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.635894 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7ltf\" (UniqueName: \"kubernetes.io/projected/03ad58cb-353a-493e-a7d0-b5c66c3261bf-kube-api-access-d7ltf\") pod \"neutron-db-sync-5jlhb\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.648833 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-xq58z"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.650242 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.653708 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.654887 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-f825x" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.677555 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xq58z"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.693980 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-nxfq4"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.704987 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c968c7f55-8j5pw"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.720281 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.730521 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.737537 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.737572 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-db-sync-config-data\") pod \"barbican-db-sync-xq58z\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.737601 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-config\") pod \"neutron-db-sync-5jlhb\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.737659 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-run-httpd\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.737690 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.737737 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-scripts\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.739196 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-combined-ca-bundle\") pod \"neutron-db-sync-5jlhb\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.739228 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-config-data\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.739249 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqb5c\" (UniqueName: \"kubernetes.io/projected/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-kube-api-access-bqb5c\") pod \"barbican-db-sync-xq58z\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.739291 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7ltf\" (UniqueName: \"kubernetes.io/projected/03ad58cb-353a-493e-a7d0-b5c66c3261bf-kube-api-access-d7ltf\") pod \"neutron-db-sync-5jlhb\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.739308 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-combined-ca-bundle\") pod \"barbican-db-sync-xq58z\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.739332 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc87w\" (UniqueName: \"kubernetes.io/projected/66d50d55-3a5f-4027-aa14-3abf8d0e3709-kube-api-access-lc87w\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.739655 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-log-httpd\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.746735 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-combined-ca-bundle\") pod \"neutron-db-sync-5jlhb\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.749671 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c968c7f55-8j5pw"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.758160 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-config\") pod \"neutron-db-sync-5jlhb\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.771949 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7ltf\" (UniqueName: \"kubernetes.io/projected/03ad58cb-353a-493e-a7d0-b5c66c3261bf-kube-api-access-d7ltf\") pod \"neutron-db-sync-5jlhb\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.780380 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-kgl2b"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.782230 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.784461 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.808168 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-kgl2b"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.842471 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqb5c\" (UniqueName: \"kubernetes.io/projected/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-kube-api-access-bqb5c\") pod \"barbican-db-sync-xq58z\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.842542 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a34147-0321-479f-92c6-d58bbf059c0b-logs\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.842585 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-dns-svc\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.842622 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-combined-ca-bundle\") pod \"barbican-db-sync-xq58z\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.843467 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc87w\" (UniqueName: \"kubernetes.io/projected/66d50d55-3a5f-4027-aa14-3abf8d0e3709-kube-api-access-lc87w\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.843580 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7rnv\" (UniqueName: \"kubernetes.io/projected/75a9acc9-8a59-4a57-abad-bed6920dcc7c-kube-api-access-h7rnv\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.843689 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-log-httpd\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.843716 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-config-data\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.843772 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-nb\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.843827 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.843854 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-db-sync-config-data\") pod \"barbican-db-sync-xq58z\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.843933 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-config\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.854473 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-run-httpd\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.859832 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-665vn"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.862695 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-combined-ca-bundle\") pod \"barbican-db-sync-xq58z\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.870831 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.887236 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqb5c\" (UniqueName: \"kubernetes.io/projected/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-kube-api-access-bqb5c\") pod \"barbican-db-sync-xq58z\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.888926 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-log-httpd\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.890149 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-run-httpd\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.894448 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-db-sync-config-data\") pod \"barbican-db-sync-xq58z\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.894827 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.894924 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0a34147-0321-479f-92c6-d58bbf059c0b-horizon-secret-key\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.895267 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvf72\" (UniqueName: \"kubernetes.io/projected/e0a34147-0321-479f-92c6-d58bbf059c0b-kube-api-access-tvf72\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.895373 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-scripts\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.895428 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-scripts\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.895529 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-sb\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.895589 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-config-data\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.901947 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-665vn"] Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.902195 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-665vn" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.906231 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-58dhc" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.910169 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.910617 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.912184 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.913826 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-scripts\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.921530 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-config-data\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.937618 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc87w\" (UniqueName: \"kubernetes.io/projected/66d50d55-3a5f-4027-aa14-3abf8d0e3709-kube-api-access-lc87w\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.940368 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " pod="openstack/ceilometer-0" Oct 09 10:22:23 crc kubenswrapper[4923]: I1009 10:22:23.951734 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.025566 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xq58z" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.026913 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-dns-svc\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027047 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7rnv\" (UniqueName: \"kubernetes.io/projected/75a9acc9-8a59-4a57-abad-bed6920dcc7c-kube-api-access-h7rnv\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027091 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-config-data\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027120 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-nb\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027158 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-scripts\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027189 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-combined-ca-bundle\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027217 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-config\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027237 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqq2d\" (UniqueName: \"kubernetes.io/projected/f250b2a7-fb69-4c40-b186-238f74472a44-kube-api-access-fqq2d\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027281 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0a34147-0321-479f-92c6-d58bbf059c0b-horizon-secret-key\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027305 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-config-data\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027336 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvf72\" (UniqueName: \"kubernetes.io/projected/e0a34147-0321-479f-92c6-d58bbf059c0b-kube-api-access-tvf72\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027356 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f250b2a7-fb69-4c40-b186-238f74472a44-logs\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027382 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-scripts\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027410 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-sb\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027433 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a34147-0321-479f-92c6-d58bbf059c0b-logs\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.027874 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a34147-0321-479f-92c6-d58bbf059c0b-logs\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.029123 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-config\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.029526 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-config-data\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.029679 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-scripts\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.032094 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-sb\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.032380 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0a34147-0321-479f-92c6-d58bbf059c0b-horizon-secret-key\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.034527 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-nb\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.035146 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-dns-svc\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.053995 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvf72\" (UniqueName: \"kubernetes.io/projected/e0a34147-0321-479f-92c6-d58bbf059c0b-kube-api-access-tvf72\") pod \"horizon-c968c7f55-8j5pw\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.057482 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7rnv\" (UniqueName: \"kubernetes.io/projected/75a9acc9-8a59-4a57-abad-bed6920dcc7c-kube-api-access-h7rnv\") pod \"dnsmasq-dns-7b99bccc6c-kgl2b\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.057602 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.117564 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.129724 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-config-data\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.129829 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f250b2a7-fb69-4c40-b186-238f74472a44-logs\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.129923 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-scripts\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.129948 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-combined-ca-bundle\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.129974 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqq2d\" (UniqueName: \"kubernetes.io/projected/f250b2a7-fb69-4c40-b186-238f74472a44-kube-api-access-fqq2d\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.130422 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f250b2a7-fb69-4c40-b186-238f74472a44-logs\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.141650 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-config-data\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.144679 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-scripts\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.148571 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-combined-ca-bundle\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.157789 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqq2d\" (UniqueName: \"kubernetes.io/projected/f250b2a7-fb69-4c40-b186-238f74472a44-kube-api-access-fqq2d\") pod \"placement-db-sync-665vn\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.193874 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-nxfq4"] Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.234877 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-665vn" Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.420717 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5fbrp"] Oct 09 10:22:24 crc kubenswrapper[4923]: W1009 10:22:24.440073 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9fb0fa45_d830_4b81_9383_39b8e622fe0c.slice/crio-395ef3b3a82fce01ee21f0ee2f2ad9774f21ee507ca67374a93ecdaa083bac55 WatchSource:0}: Error finding container 395ef3b3a82fce01ee21f0ee2f2ad9774f21ee507ca67374a93ecdaa083bac55: Status 404 returned error can't find the container with id 395ef3b3a82fce01ee21f0ee2f2ad9774f21ee507ca67374a93ecdaa083bac55 Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.494342 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-776bc5df9c-2wltl"] Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.582006 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-5ll8s"] Oct 09 10:22:24 crc kubenswrapper[4923]: W1009 10:22:24.587925 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c688fdd_7d04_4a0a_9074_94c9f1b8a5fd.slice/crio-2dfd5b2fdc6e79f2067348aad875a022677b2033adfa4be051c6ac4337c16f56 WatchSource:0}: Error finding container 2dfd5b2fdc6e79f2067348aad875a022677b2033adfa4be051c6ac4337c16f56: Status 404 returned error can't find the container with id 2dfd5b2fdc6e79f2067348aad875a022677b2033adfa4be051c6ac4337c16f56 Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.783296 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5fbrp" event={"ID":"9fb0fa45-d830-4b81-9383-39b8e622fe0c","Type":"ContainerStarted","Data":"395ef3b3a82fce01ee21f0ee2f2ad9774f21ee507ca67374a93ecdaa083bac55"} Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.786611 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" event={"ID":"c7fc5632-a21c-436c-8a74-7c99de29915d","Type":"ContainerStarted","Data":"ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909"} Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.786671 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" event={"ID":"c7fc5632-a21c-436c-8a74-7c99de29915d","Type":"ContainerStarted","Data":"384dd993964c99cde8ede8325ef1c2b971331da2779f62aac795b7c7ef42b069"} Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.788722 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776bc5df9c-2wltl" event={"ID":"75eb01a3-916a-4825-a8b4-1de9970e4e6e","Type":"ContainerStarted","Data":"feb7514ca3a8e4bf65e9ce15c458a97a794555e2300c57fc072d0805892a96fd"} Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.791045 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5ll8s" event={"ID":"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd","Type":"ContainerStarted","Data":"2dfd5b2fdc6e79f2067348aad875a022677b2033adfa4be051c6ac4337c16f56"} Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.797529 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-5jlhb"] Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.805688 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.905974 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-kgl2b"] Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.915799 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c968c7f55-8j5pw"] Oct 09 10:22:24 crc kubenswrapper[4923]: I1009 10:22:24.923172 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-xq58z"] Oct 09 10:22:24 crc kubenswrapper[4923]: W1009 10:22:24.954772 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0a34147_0321_479f_92c6_d58bbf059c0b.slice/crio-640fae7360e71750e8ad1ea7f433ef4c6b74839f97e9230adcfc13ca6da5aa26 WatchSource:0}: Error finding container 640fae7360e71750e8ad1ea7f433ef4c6b74839f97e9230adcfc13ca6da5aa26: Status 404 returned error can't find the container with id 640fae7360e71750e8ad1ea7f433ef4c6b74839f97e9230adcfc13ca6da5aa26 Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.090933 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-665vn"] Oct 09 10:22:25 crc kubenswrapper[4923]: W1009 10:22:25.105485 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf250b2a7_fb69_4c40_b186_238f74472a44.slice/crio-3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a WatchSource:0}: Error finding container 3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a: Status 404 returned error can't find the container with id 3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.201642 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.291777 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-dns-svc\") pod \"c7fc5632-a21c-436c-8a74-7c99de29915d\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.292629 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-nb\") pod \"c7fc5632-a21c-436c-8a74-7c99de29915d\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.292706 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-sb\") pod \"c7fc5632-a21c-436c-8a74-7c99de29915d\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.292779 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-config\") pod \"c7fc5632-a21c-436c-8a74-7c99de29915d\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.292875 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgbxw\" (UniqueName: \"kubernetes.io/projected/c7fc5632-a21c-436c-8a74-7c99de29915d-kube-api-access-jgbxw\") pod \"c7fc5632-a21c-436c-8a74-7c99de29915d\" (UID: \"c7fc5632-a21c-436c-8a74-7c99de29915d\") " Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.300498 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7fc5632-a21c-436c-8a74-7c99de29915d-kube-api-access-jgbxw" (OuterVolumeSpecName: "kube-api-access-jgbxw") pod "c7fc5632-a21c-436c-8a74-7c99de29915d" (UID: "c7fc5632-a21c-436c-8a74-7c99de29915d"). InnerVolumeSpecName "kube-api-access-jgbxw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.324565 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-config" (OuterVolumeSpecName: "config") pod "c7fc5632-a21c-436c-8a74-7c99de29915d" (UID: "c7fc5632-a21c-436c-8a74-7c99de29915d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.329025 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c7fc5632-a21c-436c-8a74-7c99de29915d" (UID: "c7fc5632-a21c-436c-8a74-7c99de29915d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.332194 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c7fc5632-a21c-436c-8a74-7c99de29915d" (UID: "c7fc5632-a21c-436c-8a74-7c99de29915d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.337864 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c7fc5632-a21c-436c-8a74-7c99de29915d" (UID: "c7fc5632-a21c-436c-8a74-7c99de29915d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.397528 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.397584 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.397602 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.397615 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgbxw\" (UniqueName: \"kubernetes.io/projected/c7fc5632-a21c-436c-8a74-7c99de29915d-kube-api-access-jgbxw\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.397626 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c7fc5632-a21c-436c-8a74-7c99de29915d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.563947 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-776bc5df9c-2wltl"] Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.669804 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.700106 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7c85c4b745-jf7ck"] Oct 09 10:22:25 crc kubenswrapper[4923]: E1009 10:22:25.701258 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7fc5632-a21c-436c-8a74-7c99de29915d" containerName="init" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.701283 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7fc5632-a21c-436c-8a74-7c99de29915d" containerName="init" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.701688 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7fc5632-a21c-436c-8a74-7c99de29915d" containerName="init" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.703545 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.820003 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx57v\" (UniqueName: \"kubernetes.io/projected/e39f0227-4155-4d15-9743-d81095c77c0d-kube-api-access-gx57v\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.820100 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-config-data\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.820154 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-scripts\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.820185 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e39f0227-4155-4d15-9743-d81095c77c0d-horizon-secret-key\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.820250 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e39f0227-4155-4d15-9743-d81095c77c0d-logs\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.832621 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c85c4b745-jf7ck"] Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.906369 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5jlhb" event={"ID":"03ad58cb-353a-493e-a7d0-b5c66c3261bf","Type":"ContainerStarted","Data":"be8f5935fa858eb0b33b9f22b84de796c7dab680e5ce7948cf4065b5202b3c16"} Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.907303 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5jlhb" event={"ID":"03ad58cb-353a-493e-a7d0-b5c66c3261bf","Type":"ContainerStarted","Data":"fb413e1b0b9dc46934817e57c427665e9567c516ff8fb850b263b7f630d8f809"} Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.932384 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx57v\" (UniqueName: \"kubernetes.io/projected/e39f0227-4155-4d15-9743-d81095c77c0d-kube-api-access-gx57v\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.932450 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-config-data\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.932493 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-scripts\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.932521 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e39f0227-4155-4d15-9743-d81095c77c0d-horizon-secret-key\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.932571 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e39f0227-4155-4d15-9743-d81095c77c0d-logs\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.934074 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-scripts\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.934932 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e39f0227-4155-4d15-9743-d81095c77c0d-logs\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.935189 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-665vn" event={"ID":"f250b2a7-fb69-4c40-b186-238f74472a44","Type":"ContainerStarted","Data":"3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a"} Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.935608 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-config-data\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.944976 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c968c7f55-8j5pw" event={"ID":"e0a34147-0321-479f-92c6-d58bbf059c0b","Type":"ContainerStarted","Data":"640fae7360e71750e8ad1ea7f433ef4c6b74839f97e9230adcfc13ca6da5aa26"} Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.960488 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-5jlhb" podStartSLOduration=2.960459333 podStartE2EDuration="2.960459333s" podCreationTimestamp="2025-10-09 10:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:22:25.955782763 +0000 UTC m=+1032.023964519" watchObservedRunningTime="2025-10-09 10:22:25.960459333 +0000 UTC m=+1032.028641089" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.970532 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e39f0227-4155-4d15-9743-d81095c77c0d-horizon-secret-key\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.974113 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerStarted","Data":"77b40526dd4b1429861b03227e663d440f5dc657b8378055e09ca5e89731abc4"} Oct 09 10:22:25 crc kubenswrapper[4923]: I1009 10:22:25.986108 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx57v\" (UniqueName: \"kubernetes.io/projected/e39f0227-4155-4d15-9743-d81095c77c0d-kube-api-access-gx57v\") pod \"horizon-7c85c4b745-jf7ck\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.012363 4923 generic.go:334] "Generic (PLEG): container finished" podID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" containerID="3731ed5bf853925dcee54aa6a455b4f1aadc8d102bd3eb9cea29bb1dc1f1a975" exitCode=0 Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.012511 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" event={"ID":"75a9acc9-8a59-4a57-abad-bed6920dcc7c","Type":"ContainerDied","Data":"3731ed5bf853925dcee54aa6a455b4f1aadc8d102bd3eb9cea29bb1dc1f1a975"} Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.012552 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" event={"ID":"75a9acc9-8a59-4a57-abad-bed6920dcc7c","Type":"ContainerStarted","Data":"33a2174705feb44a11cab4afd33505168f0ee05745bca9c8c4f104b6d5facba5"} Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.038643 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5fbrp" event={"ID":"9fb0fa45-d830-4b81-9383-39b8e622fe0c","Type":"ContainerStarted","Data":"3ce6673e43a0bde733503e5d3e1e3534302cb96687208e1450fb761722773acb"} Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.045867 4923 generic.go:334] "Generic (PLEG): container finished" podID="c7fc5632-a21c-436c-8a74-7c99de29915d" containerID="ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909" exitCode=0 Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.045995 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" event={"ID":"c7fc5632-a21c-436c-8a74-7c99de29915d","Type":"ContainerDied","Data":"ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909"} Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.046038 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" event={"ID":"c7fc5632-a21c-436c-8a74-7c99de29915d","Type":"ContainerDied","Data":"384dd993964c99cde8ede8325ef1c2b971331da2779f62aac795b7c7ef42b069"} Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.046062 4923 scope.go:117] "RemoveContainer" containerID="ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909" Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.046298 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67bcfd764f-nxfq4" Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.071225 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xq58z" event={"ID":"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07","Type":"ContainerStarted","Data":"d8ad35af505b36c417b15e9377d7dad77649d92bc90d5ee0bbf05f6824e50748"} Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.095187 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.114234 4923 scope.go:117] "RemoveContainer" containerID="ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909" Oct 09 10:22:26 crc kubenswrapper[4923]: E1009 10:22:26.120191 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909\": container with ID starting with ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909 not found: ID does not exist" containerID="ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909" Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.120276 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909"} err="failed to get container status \"ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909\": rpc error: code = NotFound desc = could not find container \"ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909\": container with ID starting with ba2ca91726cec7ff4730a175ca3e6b9a42849cf28a6f78a62f25cfcde5b95909 not found: ID does not exist" Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.159217 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5fbrp" podStartSLOduration=3.159180286 podStartE2EDuration="3.159180286s" podCreationTimestamp="2025-10-09 10:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:22:26.120596578 +0000 UTC m=+1032.188778334" watchObservedRunningTime="2025-10-09 10:22:26.159180286 +0000 UTC m=+1032.227362042" Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.178730 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-nxfq4"] Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.187417 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67bcfd764f-nxfq4"] Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.615228 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7fc5632-a21c-436c-8a74-7c99de29915d" path="/var/lib/kubelet/pods/c7fc5632-a21c-436c-8a74-7c99de29915d/volumes" Oct 09 10:22:26 crc kubenswrapper[4923]: I1009 10:22:26.771850 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7c85c4b745-jf7ck"] Oct 09 10:22:26 crc kubenswrapper[4923]: W1009 10:22:26.838399 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode39f0227_4155_4d15_9743_d81095c77c0d.slice/crio-9905da73f429ef7103df51c18942a7ce0a8d2b29d5462a5c92db7d4c081d2087 WatchSource:0}: Error finding container 9905da73f429ef7103df51c18942a7ce0a8d2b29d5462a5c92db7d4c081d2087: Status 404 returned error can't find the container with id 9905da73f429ef7103df51c18942a7ce0a8d2b29d5462a5c92db7d4c081d2087 Oct 09 10:22:27 crc kubenswrapper[4923]: I1009 10:22:27.096797 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c85c4b745-jf7ck" event={"ID":"e39f0227-4155-4d15-9743-d81095c77c0d","Type":"ContainerStarted","Data":"9905da73f429ef7103df51c18942a7ce0a8d2b29d5462a5c92db7d4c081d2087"} Oct 09 10:22:27 crc kubenswrapper[4923]: I1009 10:22:27.104063 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" event={"ID":"75a9acc9-8a59-4a57-abad-bed6920dcc7c","Type":"ContainerStarted","Data":"b28eb009542785a30bfc4c524f2bd23830d8e5811c093f37f918fd4660bf9ce9"} Oct 09 10:22:27 crc kubenswrapper[4923]: I1009 10:22:27.104548 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:27 crc kubenswrapper[4923]: I1009 10:22:27.134142 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" podStartSLOduration=4.134104487 podStartE2EDuration="4.134104487s" podCreationTimestamp="2025-10-09 10:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:22:27.133230903 +0000 UTC m=+1033.201412659" watchObservedRunningTime="2025-10-09 10:22:27.134104487 +0000 UTC m=+1033.202286243" Oct 09 10:22:30 crc kubenswrapper[4923]: I1009 10:22:30.154252 4923 generic.go:334] "Generic (PLEG): container finished" podID="9fb0fa45-d830-4b81-9383-39b8e622fe0c" containerID="3ce6673e43a0bde733503e5d3e1e3534302cb96687208e1450fb761722773acb" exitCode=0 Oct 09 10:22:30 crc kubenswrapper[4923]: I1009 10:22:30.154477 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5fbrp" event={"ID":"9fb0fa45-d830-4b81-9383-39b8e622fe0c","Type":"ContainerDied","Data":"3ce6673e43a0bde733503e5d3e1e3534302cb96687208e1450fb761722773acb"} Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.670670 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c968c7f55-8j5pw"] Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.742424 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-855475c586-p22m4"] Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.744767 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.758115 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.777532 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-855475c586-p22m4"] Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.875176 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c85c4b745-jf7ck"] Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.884970 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-combined-ca-bundle\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.885962 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-scripts\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.886020 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-tls-certs\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.886070 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-config-data\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.886136 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-secret-key\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.886425 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgw7x\" (UniqueName: \"kubernetes.io/projected/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-kube-api-access-cgw7x\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.886527 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-logs\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.917498 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5dd576c454-s7hpl"] Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.921522 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.946620 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5dd576c454-s7hpl"] Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988559 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-secret-key\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988622 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d75434e-89d3-464b-af06-fb0ac9f92fd8-scripts\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988700 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgw7x\" (UniqueName: \"kubernetes.io/projected/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-kube-api-access-cgw7x\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988731 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-logs\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988765 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d75434e-89d3-464b-af06-fb0ac9f92fd8-horizon-tls-certs\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988808 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d75434e-89d3-464b-af06-fb0ac9f92fd8-logs\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988856 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4d75434e-89d3-464b-af06-fb0ac9f92fd8-horizon-secret-key\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988907 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d75434e-89d3-464b-af06-fb0ac9f92fd8-combined-ca-bundle\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988932 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-combined-ca-bundle\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988953 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d75434e-89d3-464b-af06-fb0ac9f92fd8-config-data\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.988994 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgq5g\" (UniqueName: \"kubernetes.io/projected/4d75434e-89d3-464b-af06-fb0ac9f92fd8-kube-api-access-cgq5g\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.989021 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-scripts\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.989063 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-tls-certs\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.989089 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-config-data\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.989660 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-logs\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.992197 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-scripts\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.997265 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-config-data\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.997856 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-combined-ca-bundle\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.998124 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-secret-key\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:31 crc kubenswrapper[4923]: I1009 10:22:31.998407 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-tls-certs\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.011332 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgw7x\" (UniqueName: \"kubernetes.io/projected/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-kube-api-access-cgw7x\") pod \"horizon-855475c586-p22m4\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.076133 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-855475c586-p22m4" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.094737 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgq5g\" (UniqueName: \"kubernetes.io/projected/4d75434e-89d3-464b-af06-fb0ac9f92fd8-kube-api-access-cgq5g\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.094947 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d75434e-89d3-464b-af06-fb0ac9f92fd8-scripts\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.095115 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d75434e-89d3-464b-af06-fb0ac9f92fd8-horizon-tls-certs\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.095161 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d75434e-89d3-464b-af06-fb0ac9f92fd8-logs\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.095234 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4d75434e-89d3-464b-af06-fb0ac9f92fd8-horizon-secret-key\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.095955 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d75434e-89d3-464b-af06-fb0ac9f92fd8-combined-ca-bundle\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.096085 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d75434e-89d3-464b-af06-fb0ac9f92fd8-config-data\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.096343 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4d75434e-89d3-464b-af06-fb0ac9f92fd8-scripts\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.096741 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4d75434e-89d3-464b-af06-fb0ac9f92fd8-logs\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.103656 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d75434e-89d3-464b-af06-fb0ac9f92fd8-combined-ca-bundle\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.106350 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4d75434e-89d3-464b-af06-fb0ac9f92fd8-config-data\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.130181 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/4d75434e-89d3-464b-af06-fb0ac9f92fd8-horizon-secret-key\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.133383 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgq5g\" (UniqueName: \"kubernetes.io/projected/4d75434e-89d3-464b-af06-fb0ac9f92fd8-kube-api-access-cgq5g\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.153599 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d75434e-89d3-464b-af06-fb0ac9f92fd8-horizon-tls-certs\") pod \"horizon-5dd576c454-s7hpl\" (UID: \"4d75434e-89d3-464b-af06-fb0ac9f92fd8\") " pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:32 crc kubenswrapper[4923]: I1009 10:22:32.248394 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:22:34 crc kubenswrapper[4923]: I1009 10:22:34.121713 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:22:34 crc kubenswrapper[4923]: I1009 10:22:34.190746 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bblwh"] Oct 09 10:22:34 crc kubenswrapper[4923]: I1009 10:22:34.191456 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="dnsmasq-dns" containerID="cri-o://bc03f56b265729775f1504bcc57cd7ad1d38ee3f4d52e0a69914696894dd5d01" gracePeriod=10 Oct 09 10:22:34 crc kubenswrapper[4923]: I1009 10:22:34.352059 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: connect: connection refused" Oct 09 10:22:35 crc kubenswrapper[4923]: I1009 10:22:35.216324 4923 generic.go:334] "Generic (PLEG): container finished" podID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerID="bc03f56b265729775f1504bcc57cd7ad1d38ee3f4d52e0a69914696894dd5d01" exitCode=0 Oct 09 10:22:35 crc kubenswrapper[4923]: I1009 10:22:35.216408 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" event={"ID":"255708b2-d7b2-4f19-9ae2-5cf29d005025","Type":"ContainerDied","Data":"bc03f56b265729775f1504bcc57cd7ad1d38ee3f4d52e0a69914696894dd5d01"} Oct 09 10:22:39 crc kubenswrapper[4923]: E1009 10:22:39.594271 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:59448516174fc3bab679b9a8dd62cb9a9d16b5734aadbeb98e960e3b7c79bd22" Oct 09 10:22:39 crc kubenswrapper[4923]: E1009 10:22:39.595107 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:59448516174fc3bab679b9a8dd62cb9a9d16b5734aadbeb98e960e3b7c79bd22,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fqq2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-665vn_openstack(f250b2a7-fb69-4c40-b186-238f74472a44): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:22:39 crc kubenswrapper[4923]: E1009 10:22:39.596387 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-665vn" podUID="f250b2a7-fb69-4c40-b186-238f74472a44" Oct 09 10:22:40 crc kubenswrapper[4923]: E1009 10:22:40.275181 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:59448516174fc3bab679b9a8dd62cb9a9d16b5734aadbeb98e960e3b7c79bd22\\\"\"" pod="openstack/placement-db-sync-665vn" podUID="f250b2a7-fb69-4c40-b186-238f74472a44" Oct 09 10:22:42 crc kubenswrapper[4923]: E1009 10:22:41.343971 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490" Oct 09 10:22:42 crc kubenswrapper[4923]: E1009 10:22:41.344541 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d5h67fh5f9hbhc9h685h55fhb9h5fchf6h58bh5bfhc4h699h9bh78h655h78h587h577h95h666h5b7hbfh58dh67bh645h5ch58dh5f4hc8h65bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wszh4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-776bc5df9c-2wltl_openstack(75eb01a3-916a-4825-a8b4-1de9970e4e6e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:22:42 crc kubenswrapper[4923]: E1009 10:22:41.351473 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490\\\"\"]" pod="openstack/horizon-776bc5df9c-2wltl" podUID="75eb01a3-916a-4825-a8b4-1de9970e4e6e" Oct 09 10:22:42 crc kubenswrapper[4923]: E1009 10:22:41.352265 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490" Oct 09 10:22:42 crc kubenswrapper[4923]: E1009 10:22:41.352474 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5cbh8h574h68h556h6ch694h679hb9h657h648h5b7h697h98h79hch64dh69hb9h65ch58h675h8ch579h595h5fch55bhdh8h5f6hdch575q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tvf72,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-c968c7f55-8j5pw_openstack(e0a34147-0321-479f-92c6-d58bbf059c0b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:22:42 crc kubenswrapper[4923]: E1009 10:22:41.362954 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:03b4bb79b71d5ca7792d19c4c0ee08a5e5a407ad844c087305c42dd909ee7490\\\"\"]" pod="openstack/horizon-c968c7f55-8j5pw" podUID="e0a34147-0321-479f-92c6-d58bbf059c0b" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.399359 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.515280 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-combined-ca-bundle\") pod \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.515337 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4mwc\" (UniqueName: \"kubernetes.io/projected/9fb0fa45-d830-4b81-9383-39b8e622fe0c-kube-api-access-f4mwc\") pod \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.515377 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-credential-keys\") pod \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.515430 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-fernet-keys\") pod \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.515509 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-scripts\") pod \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.515571 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-config-data\") pod \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\" (UID: \"9fb0fa45-d830-4b81-9383-39b8e622fe0c\") " Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.528957 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-scripts" (OuterVolumeSpecName: "scripts") pod "9fb0fa45-d830-4b81-9383-39b8e622fe0c" (UID: "9fb0fa45-d830-4b81-9383-39b8e622fe0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.529019 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "9fb0fa45-d830-4b81-9383-39b8e622fe0c" (UID: "9fb0fa45-d830-4b81-9383-39b8e622fe0c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.529061 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "9fb0fa45-d830-4b81-9383-39b8e622fe0c" (UID: "9fb0fa45-d830-4b81-9383-39b8e622fe0c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.529118 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fb0fa45-d830-4b81-9383-39b8e622fe0c-kube-api-access-f4mwc" (OuterVolumeSpecName: "kube-api-access-f4mwc") pod "9fb0fa45-d830-4b81-9383-39b8e622fe0c" (UID: "9fb0fa45-d830-4b81-9383-39b8e622fe0c"). InnerVolumeSpecName "kube-api-access-f4mwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.548510 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9fb0fa45-d830-4b81-9383-39b8e622fe0c" (UID: "9fb0fa45-d830-4b81-9383-39b8e622fe0c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.566661 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-config-data" (OuterVolumeSpecName: "config-data") pod "9fb0fa45-d830-4b81-9383-39b8e622fe0c" (UID: "9fb0fa45-d830-4b81-9383-39b8e622fe0c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.618303 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.618338 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4mwc\" (UniqueName: \"kubernetes.io/projected/9fb0fa45-d830-4b81-9383-39b8e622fe0c-kube-api-access-f4mwc\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.618352 4923 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.618362 4923 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.618371 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:41.618382 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9fb0fa45-d830-4b81-9383-39b8e622fe0c-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.290931 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5fbrp" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.300004 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5fbrp" event={"ID":"9fb0fa45-d830-4b81-9383-39b8e622fe0c","Type":"ContainerDied","Data":"395ef3b3a82fce01ee21f0ee2f2ad9774f21ee507ca67374a93ecdaa083bac55"} Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.300063 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="395ef3b3a82fce01ee21f0ee2f2ad9774f21ee507ca67374a93ecdaa083bac55" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.488597 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5fbrp"] Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.495312 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5fbrp"] Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.592639 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2m7bb"] Oct 09 10:22:42 crc kubenswrapper[4923]: E1009 10:22:42.594014 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fb0fa45-d830-4b81-9383-39b8e622fe0c" containerName="keystone-bootstrap" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.594039 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fb0fa45-d830-4b81-9383-39b8e622fe0c" containerName="keystone-bootstrap" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.594390 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fb0fa45-d830-4b81-9383-39b8e622fe0c" containerName="keystone-bootstrap" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.599773 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.604650 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.604734 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.604980 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-58pgm" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.609102 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.671906 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fb0fa45-d830-4b81-9383-39b8e622fe0c" path="/var/lib/kubelet/pods/9fb0fa45-d830-4b81-9383-39b8e622fe0c/volumes" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.677285 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2m7bb"] Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.751263 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-credential-keys\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.751321 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-combined-ca-bundle\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.751352 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvp2r\" (UniqueName: \"kubernetes.io/projected/85c95222-efd5-4bd6-ab01-c51eb21abb4c-kube-api-access-cvp2r\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.751450 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-fernet-keys\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.752568 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-config-data\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.752779 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-scripts\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.855518 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-fernet-keys\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.855635 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-config-data\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.855773 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-scripts\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.855828 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-credential-keys\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.855868 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-combined-ca-bundle\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.855904 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvp2r\" (UniqueName: \"kubernetes.io/projected/85c95222-efd5-4bd6-ab01-c51eb21abb4c-kube-api-access-cvp2r\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.863845 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-credential-keys\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.863908 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-fernet-keys\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.864710 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-config-data\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.865829 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-scripts\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.866237 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-combined-ca-bundle\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.874921 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvp2r\" (UniqueName: \"kubernetes.io/projected/85c95222-efd5-4bd6-ab01-c51eb21abb4c-kube-api-access-cvp2r\") pod \"keystone-bootstrap-2m7bb\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:42 crc kubenswrapper[4923]: I1009 10:22:42.928720 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:22:44 crc kubenswrapper[4923]: I1009 10:22:44.351072 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Oct 09 10:22:46 crc kubenswrapper[4923]: I1009 10:22:46.331334 4923 generic.go:334] "Generic (PLEG): container finished" podID="03ad58cb-353a-493e-a7d0-b5c66c3261bf" containerID="be8f5935fa858eb0b33b9f22b84de796c7dab680e5ce7948cf4065b5202b3c16" exitCode=0 Oct 09 10:22:46 crc kubenswrapper[4923]: I1009 10:22:46.331454 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5jlhb" event={"ID":"03ad58cb-353a-493e-a7d0-b5c66c3261bf","Type":"ContainerDied","Data":"be8f5935fa858eb0b33b9f22b84de796c7dab680e5ce7948cf4065b5202b3c16"} Oct 09 10:22:49 crc kubenswrapper[4923]: I1009 10:22:49.361185 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Oct 09 10:22:49 crc kubenswrapper[4923]: I1009 10:22:49.362197 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.357869 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.407106 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" event={"ID":"255708b2-d7b2-4f19-9ae2-5cf29d005025","Type":"ContainerDied","Data":"6d1b53d8fb47f058a8168e207ec84a86ef40b240a3c36bd0d6945bfd7d17bc8d"} Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.407184 4923 scope.go:117] "RemoveContainer" containerID="bc03f56b265729775f1504bcc57cd7ad1d38ee3f4d52e0a69914696894dd5d01" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.407189 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.521818 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-dns-svc\") pod \"255708b2-d7b2-4f19-9ae2-5cf29d005025\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.522020 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-nb\") pod \"255708b2-d7b2-4f19-9ae2-5cf29d005025\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.522057 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-sb\") pod \"255708b2-d7b2-4f19-9ae2-5cf29d005025\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.522127 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-config\") pod \"255708b2-d7b2-4f19-9ae2-5cf29d005025\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.522206 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh7c5\" (UniqueName: \"kubernetes.io/projected/255708b2-d7b2-4f19-9ae2-5cf29d005025-kube-api-access-rh7c5\") pod \"255708b2-d7b2-4f19-9ae2-5cf29d005025\" (UID: \"255708b2-d7b2-4f19-9ae2-5cf29d005025\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.533238 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/255708b2-d7b2-4f19-9ae2-5cf29d005025-kube-api-access-rh7c5" (OuterVolumeSpecName: "kube-api-access-rh7c5") pod "255708b2-d7b2-4f19-9ae2-5cf29d005025" (UID: "255708b2-d7b2-4f19-9ae2-5cf29d005025"). InnerVolumeSpecName "kube-api-access-rh7c5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.571962 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "255708b2-d7b2-4f19-9ae2-5cf29d005025" (UID: "255708b2-d7b2-4f19-9ae2-5cf29d005025"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.577395 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-config" (OuterVolumeSpecName: "config") pod "255708b2-d7b2-4f19-9ae2-5cf29d005025" (UID: "255708b2-d7b2-4f19-9ae2-5cf29d005025"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.578054 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "255708b2-d7b2-4f19-9ae2-5cf29d005025" (UID: "255708b2-d7b2-4f19-9ae2-5cf29d005025"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.578931 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "255708b2-d7b2-4f19-9ae2-5cf29d005025" (UID: "255708b2-d7b2-4f19-9ae2-5cf29d005025"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.624514 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.624561 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rh7c5\" (UniqueName: \"kubernetes.io/projected/255708b2-d7b2-4f19-9ae2-5cf29d005025-kube-api-access-rh7c5\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.624577 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.624590 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.624603 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/255708b2-d7b2-4f19-9ae2-5cf29d005025-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.750645 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bblwh"] Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.765343 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74b7749bc7-bblwh"] Oct 09 10:22:53 crc kubenswrapper[4923]: E1009 10:22:53.843020 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384" Oct 09 10:22:53 crc kubenswrapper[4923]: E1009 10:22:53.843235 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bqb5c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-xq58z_openstack(31f10461-abd0-4c4d-bdb5-3da1f1e6ed07): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:22:53 crc kubenswrapper[4923]: E1009 10:22:53.845184 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-xq58z" podUID="31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.852209 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.857596 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.868006 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.932776 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-scripts\") pod \"e0a34147-0321-479f-92c6-d58bbf059c0b\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.932893 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-config-data\") pod \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.932972 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75eb01a3-916a-4825-a8b4-1de9970e4e6e-logs\") pod \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.933079 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wszh4\" (UniqueName: \"kubernetes.io/projected/75eb01a3-916a-4825-a8b4-1de9970e4e6e-kube-api-access-wszh4\") pod \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.933914 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-config-data\") pod \"e0a34147-0321-479f-92c6-d58bbf059c0b\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.934350 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0a34147-0321-479f-92c6-d58bbf059c0b-horizon-secret-key\") pod \"e0a34147-0321-479f-92c6-d58bbf059c0b\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.934451 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvf72\" (UniqueName: \"kubernetes.io/projected/e0a34147-0321-479f-92c6-d58bbf059c0b-kube-api-access-tvf72\") pod \"e0a34147-0321-479f-92c6-d58bbf059c0b\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.934539 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a34147-0321-479f-92c6-d58bbf059c0b-logs\") pod \"e0a34147-0321-479f-92c6-d58bbf059c0b\" (UID: \"e0a34147-0321-479f-92c6-d58bbf059c0b\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.934617 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75eb01a3-916a-4825-a8b4-1de9970e4e6e-horizon-secret-key\") pod \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.934804 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-scripts\") pod \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\" (UID: \"75eb01a3-916a-4825-a8b4-1de9970e4e6e\") " Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.935988 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-scripts" (OuterVolumeSpecName: "scripts") pod "75eb01a3-916a-4825-a8b4-1de9970e4e6e" (UID: "75eb01a3-916a-4825-a8b4-1de9970e4e6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.936362 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75eb01a3-916a-4825-a8b4-1de9970e4e6e-logs" (OuterVolumeSpecName: "logs") pod "75eb01a3-916a-4825-a8b4-1de9970e4e6e" (UID: "75eb01a3-916a-4825-a8b4-1de9970e4e6e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.936777 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-scripts" (OuterVolumeSpecName: "scripts") pod "e0a34147-0321-479f-92c6-d58bbf059c0b" (UID: "e0a34147-0321-479f-92c6-d58bbf059c0b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.936962 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-config-data" (OuterVolumeSpecName: "config-data") pod "e0a34147-0321-479f-92c6-d58bbf059c0b" (UID: "e0a34147-0321-479f-92c6-d58bbf059c0b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.936959 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-config-data" (OuterVolumeSpecName: "config-data") pod "75eb01a3-916a-4825-a8b4-1de9970e4e6e" (UID: "75eb01a3-916a-4825-a8b4-1de9970e4e6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.937203 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0a34147-0321-479f-92c6-d58bbf059c0b-logs" (OuterVolumeSpecName: "logs") pod "e0a34147-0321-479f-92c6-d58bbf059c0b" (UID: "e0a34147-0321-479f-92c6-d58bbf059c0b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.941297 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0a34147-0321-479f-92c6-d58bbf059c0b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e0a34147-0321-479f-92c6-d58bbf059c0b" (UID: "e0a34147-0321-479f-92c6-d58bbf059c0b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.942110 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0a34147-0321-479f-92c6-d58bbf059c0b-kube-api-access-tvf72" (OuterVolumeSpecName: "kube-api-access-tvf72") pod "e0a34147-0321-479f-92c6-d58bbf059c0b" (UID: "e0a34147-0321-479f-92c6-d58bbf059c0b"). InnerVolumeSpecName "kube-api-access-tvf72". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.942343 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75eb01a3-916a-4825-a8b4-1de9970e4e6e-kube-api-access-wszh4" (OuterVolumeSpecName: "kube-api-access-wszh4") pod "75eb01a3-916a-4825-a8b4-1de9970e4e6e" (UID: "75eb01a3-916a-4825-a8b4-1de9970e4e6e"). InnerVolumeSpecName "kube-api-access-wszh4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:53 crc kubenswrapper[4923]: I1009 10:22:53.944000 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75eb01a3-916a-4825-a8b4-1de9970e4e6e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "75eb01a3-916a-4825-a8b4-1de9970e4e6e" (UID: "75eb01a3-916a-4825-a8b4-1de9970e4e6e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.036747 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7ltf\" (UniqueName: \"kubernetes.io/projected/03ad58cb-353a-493e-a7d0-b5c66c3261bf-kube-api-access-d7ltf\") pod \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.036837 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-combined-ca-bundle\") pod \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.036905 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-config\") pod \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\" (UID: \"03ad58cb-353a-493e-a7d0-b5c66c3261bf\") " Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037510 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037526 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037535 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/75eb01a3-916a-4825-a8b4-1de9970e4e6e-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037544 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75eb01a3-916a-4825-a8b4-1de9970e4e6e-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037553 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wszh4\" (UniqueName: \"kubernetes.io/projected/75eb01a3-916a-4825-a8b4-1de9970e4e6e-kube-api-access-wszh4\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037562 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e0a34147-0321-479f-92c6-d58bbf059c0b-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037570 4923 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e0a34147-0321-479f-92c6-d58bbf059c0b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037578 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvf72\" (UniqueName: \"kubernetes.io/projected/e0a34147-0321-479f-92c6-d58bbf059c0b-kube-api-access-tvf72\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037587 4923 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/75eb01a3-916a-4825-a8b4-1de9970e4e6e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.037595 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e0a34147-0321-479f-92c6-d58bbf059c0b-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.044125 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03ad58cb-353a-493e-a7d0-b5c66c3261bf-kube-api-access-d7ltf" (OuterVolumeSpecName: "kube-api-access-d7ltf") pod "03ad58cb-353a-493e-a7d0-b5c66c3261bf" (UID: "03ad58cb-353a-493e-a7d0-b5c66c3261bf"). InnerVolumeSpecName "kube-api-access-d7ltf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.064116 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03ad58cb-353a-493e-a7d0-b5c66c3261bf" (UID: "03ad58cb-353a-493e-a7d0-b5c66c3261bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.067003 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-config" (OuterVolumeSpecName: "config") pod "03ad58cb-353a-493e-a7d0-b5c66c3261bf" (UID: "03ad58cb-353a-493e-a7d0-b5c66c3261bf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.140678 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7ltf\" (UniqueName: \"kubernetes.io/projected/03ad58cb-353a-493e-a7d0-b5c66c3261bf-kube-api-access-d7ltf\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.140729 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.140763 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/03ad58cb-353a-493e-a7d0-b5c66c3261bf-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.362386 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74b7749bc7-bblwh" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.126:5353: i/o timeout" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.417512 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-776bc5df9c-2wltl" event={"ID":"75eb01a3-916a-4825-a8b4-1de9970e4e6e","Type":"ContainerDied","Data":"feb7514ca3a8e4bf65e9ce15c458a97a794555e2300c57fc072d0805892a96fd"} Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.417528 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-776bc5df9c-2wltl" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.420407 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-5jlhb" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.420421 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-5jlhb" event={"ID":"03ad58cb-353a-493e-a7d0-b5c66c3261bf","Type":"ContainerDied","Data":"fb413e1b0b9dc46934817e57c427665e9567c516ff8fb850b263b7f630d8f809"} Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.420471 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb413e1b0b9dc46934817e57c427665e9567c516ff8fb850b263b7f630d8f809" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.422213 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c968c7f55-8j5pw" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.422247 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c968c7f55-8j5pw" event={"ID":"e0a34147-0321-479f-92c6-d58bbf059c0b","Type":"ContainerDied","Data":"640fae7360e71750e8ad1ea7f433ef4c6b74839f97e9230adcfc13ca6da5aa26"} Oct 09 10:22:54 crc kubenswrapper[4923]: E1009 10:22:54.427864 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:cbe345acb37e57986ecf6685d28c72d0e639bdb493a18e9d3ba947d6c3a16384\\\"\"" pod="openstack/barbican-db-sync-xq58z" podUID="31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.527583 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-776bc5df9c-2wltl"] Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.541636 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-776bc5df9c-2wltl"] Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.556616 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c968c7f55-8j5pw"] Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.562900 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c968c7f55-8j5pw"] Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.612523 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" path="/var/lib/kubelet/pods/255708b2-d7b2-4f19-9ae2-5cf29d005025/volumes" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.613461 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75eb01a3-916a-4825-a8b4-1de9970e4e6e" path="/var/lib/kubelet/pods/75eb01a3-916a-4825-a8b4-1de9970e4e6e/volumes" Oct 09 10:22:54 crc kubenswrapper[4923]: I1009 10:22:54.613973 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0a34147-0321-479f-92c6-d58bbf059c0b" path="/var/lib/kubelet/pods/e0a34147-0321-479f-92c6-d58bbf059c0b/volumes" Oct 09 10:22:55 crc kubenswrapper[4923]: E1009 10:22:55.008572 4923 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f" Oct 09 10:22:55 crc kubenswrapper[4923]: E1009 10:22:55.009184 4923 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r7rgc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-5ll8s_openstack(0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 10:22:55 crc kubenswrapper[4923]: E1009 10:22:55.011055 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-5ll8s" podUID="0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.014487 4923 scope.go:117] "RemoveContainer" containerID="73e71e1a63690f76fc38a9e0bcbc28a007f2b6c091a88d0b097f12ced24139c7" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.093331 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh"] Oct 09 10:22:55 crc kubenswrapper[4923]: E1009 10:22:55.093912 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="init" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.094072 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="init" Oct 09 10:22:55 crc kubenswrapper[4923]: E1009 10:22:55.094155 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03ad58cb-353a-493e-a7d0-b5c66c3261bf" containerName="neutron-db-sync" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.094207 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="03ad58cb-353a-493e-a7d0-b5c66c3261bf" containerName="neutron-db-sync" Oct 09 10:22:55 crc kubenswrapper[4923]: E1009 10:22:55.094267 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="dnsmasq-dns" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.094319 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="dnsmasq-dns" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.094553 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="255708b2-d7b2-4f19-9ae2-5cf29d005025" containerName="dnsmasq-dns" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.094620 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="03ad58cb-353a-493e-a7d0-b5c66c3261bf" containerName="neutron-db-sync" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.096744 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.114964 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh"] Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.276978 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-config\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.277507 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-sb\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.277578 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-dns-svc\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.277788 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp6cg\" (UniqueName: \"kubernetes.io/projected/8395d8be-e7bb-4136-a47a-e61770e52b99-kube-api-access-zp6cg\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.277910 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-nb\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.314851 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-665d85ccb-zkkd2"] Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.316474 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.343061 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.344773 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-c5hbk" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.344878 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.344997 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.380204 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-nb\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.380285 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-config\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.380348 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-sb\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.380396 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-dns-svc\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.380521 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp6cg\" (UniqueName: \"kubernetes.io/projected/8395d8be-e7bb-4136-a47a-e61770e52b99-kube-api-access-zp6cg\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.383414 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-nb\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.385663 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-dns-svc\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.388661 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-sb\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.394952 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-665d85ccb-zkkd2"] Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.400890 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-config\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.448663 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp6cg\" (UniqueName: \"kubernetes.io/projected/8395d8be-e7bb-4136-a47a-e61770e52b99-kube-api-access-zp6cg\") pod \"dnsmasq-dns-7f7d8dc7ff-4jtmh\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: E1009 10:22:55.474016 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:85c75d60e1bd2f8a9ea0a2bb21a8df64c0a6f7b504cc1a05a355981d4b90e92f\\\"\"" pod="openstack/cinder-db-sync-5ll8s" podUID="0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.486918 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-ovndb-tls-certs\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.487090 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-combined-ca-bundle\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.487147 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-config\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.487193 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9brq2\" (UniqueName: \"kubernetes.io/projected/f9c24b4f-b830-494a-9789-745e176ffac6-kube-api-access-9brq2\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.487320 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-httpd-config\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.531344 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.597865 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-ovndb-tls-certs\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.598134 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-combined-ca-bundle\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.599542 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-config\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.599605 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9brq2\" (UniqueName: \"kubernetes.io/projected/f9c24b4f-b830-494a-9789-745e176ffac6-kube-api-access-9brq2\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.599829 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-httpd-config\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.624160 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-httpd-config\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.631251 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-combined-ca-bundle\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.640738 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-config\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.642046 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9brq2\" (UniqueName: \"kubernetes.io/projected/f9c24b4f-b830-494a-9789-745e176ffac6-kube-api-access-9brq2\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.663992 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-ovndb-tls-certs\") pod \"neutron-665d85ccb-zkkd2\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.693302 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.833732 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-855475c586-p22m4"] Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.901416 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5dd576c454-s7hpl"] Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.958426 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2m7bb"] Oct 09 10:22:55 crc kubenswrapper[4923]: I1009 10:22:55.990184 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh"] Oct 09 10:22:56 crc kubenswrapper[4923]: W1009 10:22:56.016905 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8395d8be_e7bb_4136_a47a_e61770e52b99.slice/crio-053e3e68ab8f57d185a55cb5951c4aec6f941420cffcbff32152b62dd8b2b637 WatchSource:0}: Error finding container 053e3e68ab8f57d185a55cb5951c4aec6f941420cffcbff32152b62dd8b2b637: Status 404 returned error can't find the container with id 053e3e68ab8f57d185a55cb5951c4aec6f941420cffcbff32152b62dd8b2b637 Oct 09 10:22:56 crc kubenswrapper[4923]: W1009 10:22:56.431744 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9c24b4f_b830_494a_9789_745e176ffac6.slice/crio-5474096c79a28be441c85f6a1180a8a74a7885f18447189d15cbb555cb9b5292 WatchSource:0}: Error finding container 5474096c79a28be441c85f6a1180a8a74a7885f18447189d15cbb555cb9b5292: Status 404 returned error can't find the container with id 5474096c79a28be441c85f6a1180a8a74a7885f18447189d15cbb555cb9b5292 Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.434192 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-665d85ccb-zkkd2"] Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.503144 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dd576c454-s7hpl" event={"ID":"4d75434e-89d3-464b-af06-fb0ac9f92fd8","Type":"ContainerStarted","Data":"5fb3e7f82b8d731bb06d050b3a5ae055471b9438ff0065db81f4a31cb33d1eb9"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.503201 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dd576c454-s7hpl" event={"ID":"4d75434e-89d3-464b-af06-fb0ac9f92fd8","Type":"ContainerStarted","Data":"6eef3a9ce078fecc67995be9305d1c22db3eb3a3ffa8620193517d66360b5357"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.522023 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855475c586-p22m4" event={"ID":"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b","Type":"ContainerStarted","Data":"fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.522363 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855475c586-p22m4" event={"ID":"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b","Type":"ContainerStarted","Data":"afc3d7870ca910ccc67dce336fedeb9a97c4eba98f58587805294702f948d453"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.539949 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-665vn" event={"ID":"f250b2a7-fb69-4c40-b186-238f74472a44","Type":"ContainerStarted","Data":"8917a51a7c0f6b85a9257086102be6807938116cc331605350a66bcfa7d2e5a0"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.552425 4923 generic.go:334] "Generic (PLEG): container finished" podID="8395d8be-e7bb-4136-a47a-e61770e52b99" containerID="4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b" exitCode=0 Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.552544 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" event={"ID":"8395d8be-e7bb-4136-a47a-e61770e52b99","Type":"ContainerDied","Data":"4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.552589 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" event={"ID":"8395d8be-e7bb-4136-a47a-e61770e52b99","Type":"ContainerStarted","Data":"053e3e68ab8f57d185a55cb5951c4aec6f941420cffcbff32152b62dd8b2b637"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.618271 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-665vn" podStartSLOduration=3.502627728 podStartE2EDuration="33.618239289s" podCreationTimestamp="2025-10-09 10:22:23 +0000 UTC" firstStartedPulling="2025-10-09 10:22:25.107906681 +0000 UTC m=+1031.176088437" lastFinishedPulling="2025-10-09 10:22:55.223518242 +0000 UTC m=+1061.291699998" observedRunningTime="2025-10-09 10:22:56.602227395 +0000 UTC m=+1062.670409161" watchObservedRunningTime="2025-10-09 10:22:56.618239289 +0000 UTC m=+1062.686421045" Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.634302 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerStarted","Data":"b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.634367 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-665d85ccb-zkkd2" event={"ID":"f9c24b4f-b830-494a-9789-745e176ffac6","Type":"ContainerStarted","Data":"5474096c79a28be441c85f6a1180a8a74a7885f18447189d15cbb555cb9b5292"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.670207 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2m7bb" event={"ID":"85c95222-efd5-4bd6-ab01-c51eb21abb4c","Type":"ContainerStarted","Data":"379ad10c6f26658939e5b9c00ad61fcc48f49cba5639a288227a541078da67cf"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.670277 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2m7bb" event={"ID":"85c95222-efd5-4bd6-ab01-c51eb21abb4c","Type":"ContainerStarted","Data":"b060555b0665e34cfc8d1a492ad44bbf786abc91a225bf2d3b6ca1e8f38a400a"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.698298 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c85c4b745-jf7ck" event={"ID":"e39f0227-4155-4d15-9743-d81095c77c0d","Type":"ContainerStarted","Data":"146a9d63b4b576ce6c37a77aa8c2b45c7a68cf6c37f54f1136ced8d817fc95d8"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.698475 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c85c4b745-jf7ck" event={"ID":"e39f0227-4155-4d15-9743-d81095c77c0d","Type":"ContainerStarted","Data":"230fb6ba9b99d46c0ba822b55f186e30ac29a85fef94e58740f5ad27d43191a3"} Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.698735 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c85c4b745-jf7ck" podUID="e39f0227-4155-4d15-9743-d81095c77c0d" containerName="horizon-log" containerID="cri-o://230fb6ba9b99d46c0ba822b55f186e30ac29a85fef94e58740f5ad27d43191a3" gracePeriod=30 Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.698940 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7c85c4b745-jf7ck" podUID="e39f0227-4155-4d15-9743-d81095c77c0d" containerName="horizon" containerID="cri-o://146a9d63b4b576ce6c37a77aa8c2b45c7a68cf6c37f54f1136ced8d817fc95d8" gracePeriod=30 Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.704203 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2m7bb" podStartSLOduration=14.704182139 podStartE2EDuration="14.704182139s" podCreationTimestamp="2025-10-09 10:22:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:22:56.70352095 +0000 UTC m=+1062.771702706" watchObservedRunningTime="2025-10-09 10:22:56.704182139 +0000 UTC m=+1062.772363885" Oct 09 10:22:56 crc kubenswrapper[4923]: I1009 10:22:56.752385 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7c85c4b745-jf7ck" podStartSLOduration=3.641432033 podStartE2EDuration="31.752348133s" podCreationTimestamp="2025-10-09 10:22:25 +0000 UTC" firstStartedPulling="2025-10-09 10:22:26.850550684 +0000 UTC m=+1032.918732430" lastFinishedPulling="2025-10-09 10:22:54.961466774 +0000 UTC m=+1061.029648530" observedRunningTime="2025-10-09 10:22:56.745240926 +0000 UTC m=+1062.813422692" watchObservedRunningTime="2025-10-09 10:22:56.752348133 +0000 UTC m=+1062.820529889" Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.710854 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855475c586-p22m4" event={"ID":"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b","Type":"ContainerStarted","Data":"b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667"} Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.713232 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dd576c454-s7hpl" event={"ID":"4d75434e-89d3-464b-af06-fb0ac9f92fd8","Type":"ContainerStarted","Data":"c9022b93edc7ac16068607a6f0019cf0d345890be02fad69fba4cd3fee50126a"} Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.716049 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" event={"ID":"8395d8be-e7bb-4136-a47a-e61770e52b99","Type":"ContainerStarted","Data":"36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1"} Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.716626 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.719269 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-665d85ccb-zkkd2" event={"ID":"f9c24b4f-b830-494a-9789-745e176ffac6","Type":"ContainerStarted","Data":"86bf1e0b381981b4936ffb857b09e0bfb1091b6d8d9c1a65c1150d5f4cfaa492"} Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.719302 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.719318 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-665d85ccb-zkkd2" event={"ID":"f9c24b4f-b830-494a-9789-745e176ffac6","Type":"ContainerStarted","Data":"b95dac3af77883b7de07aef6630527ed09e7cfba4c7092318dd041d2804ba167"} Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.747924 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-855475c586-p22m4" podStartSLOduration=26.747889404 podStartE2EDuration="26.747889404s" podCreationTimestamp="2025-10-09 10:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:22:57.745890089 +0000 UTC m=+1063.814071865" watchObservedRunningTime="2025-10-09 10:22:57.747889404 +0000 UTC m=+1063.816071160" Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.805013 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5dd576c454-s7hpl" podStartSLOduration=26.804989646 podStartE2EDuration="26.804989646s" podCreationTimestamp="2025-10-09 10:22:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:22:57.801249752 +0000 UTC m=+1063.869431508" watchObservedRunningTime="2025-10-09 10:22:57.804989646 +0000 UTC m=+1063.873171402" Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.809303 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" podStartSLOduration=2.809280765 podStartE2EDuration="2.809280765s" podCreationTimestamp="2025-10-09 10:22:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:22:57.769799881 +0000 UTC m=+1063.837981647" watchObservedRunningTime="2025-10-09 10:22:57.809280765 +0000 UTC m=+1063.877462541" Oct 09 10:22:57 crc kubenswrapper[4923]: I1009 10:22:57.835692 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-665d85ccb-zkkd2" podStartSLOduration=2.8356689250000002 podStartE2EDuration="2.835668925s" podCreationTimestamp="2025-10-09 10:22:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:22:57.823044456 +0000 UTC m=+1063.891226212" watchObservedRunningTime="2025-10-09 10:22:57.835668925 +0000 UTC m=+1063.903850681" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.302573 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5fbfdd98cc-7gf4c"] Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.304791 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.308146 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.319045 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.326815 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fbfdd98cc-7gf4c"] Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.401466 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-httpd-config\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.401541 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-szm24\" (UniqueName: \"kubernetes.io/projected/ac145f4f-e2db-4f2f-8842-7256514a4ee2-kube-api-access-szm24\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.401580 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-config\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.401717 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-ovndb-tls-certs\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.401806 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-internal-tls-certs\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.401946 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-combined-ca-bundle\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.402322 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-public-tls-certs\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.504194 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-httpd-config\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.504257 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-szm24\" (UniqueName: \"kubernetes.io/projected/ac145f4f-e2db-4f2f-8842-7256514a4ee2-kube-api-access-szm24\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.504294 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-config\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.504358 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-ovndb-tls-certs\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.504411 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-internal-tls-certs\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.504488 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-combined-ca-bundle\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.504582 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-public-tls-certs\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.514627 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-public-tls-certs\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.516769 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-httpd-config\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.517487 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-ovndb-tls-certs\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.518026 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-config\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.518080 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-internal-tls-certs\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.520489 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac145f4f-e2db-4f2f-8842-7256514a4ee2-combined-ca-bundle\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.523984 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-szm24\" (UniqueName: \"kubernetes.io/projected/ac145f4f-e2db-4f2f-8842-7256514a4ee2-kube-api-access-szm24\") pod \"neutron-5fbfdd98cc-7gf4c\" (UID: \"ac145f4f-e2db-4f2f-8842-7256514a4ee2\") " pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:58 crc kubenswrapper[4923]: I1009 10:22:58.648606 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:22:59 crc kubenswrapper[4923]: I1009 10:22:59.323300 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5fbfdd98cc-7gf4c"] Oct 09 10:22:59 crc kubenswrapper[4923]: I1009 10:22:59.772937 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbfdd98cc-7gf4c" event={"ID":"ac145f4f-e2db-4f2f-8842-7256514a4ee2","Type":"ContainerStarted","Data":"519a31873705853d940254e9e7ed769a14d2399d0688d16be331b8a843b29128"} Oct 09 10:23:02 crc kubenswrapper[4923]: I1009 10:23:02.076255 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-855475c586-p22m4" Oct 09 10:23:02 crc kubenswrapper[4923]: I1009 10:23:02.079203 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-855475c586-p22m4" Oct 09 10:23:02 crc kubenswrapper[4923]: I1009 10:23:02.249363 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:23:02 crc kubenswrapper[4923]: I1009 10:23:02.250845 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:23:05 crc kubenswrapper[4923]: I1009 10:23:05.535013 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:23:05 crc kubenswrapper[4923]: I1009 10:23:05.600863 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-kgl2b"] Oct 09 10:23:05 crc kubenswrapper[4923]: I1009 10:23:05.601397 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" podUID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" containerName="dnsmasq-dns" containerID="cri-o://b28eb009542785a30bfc4c524f2bd23830d8e5811c093f37f918fd4660bf9ce9" gracePeriod=10 Oct 09 10:23:06 crc kubenswrapper[4923]: I1009 10:23:06.096797 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:23:07 crc kubenswrapper[4923]: I1009 10:23:07.864901 4923 generic.go:334] "Generic (PLEG): container finished" podID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" containerID="b28eb009542785a30bfc4c524f2bd23830d8e5811c093f37f918fd4660bf9ce9" exitCode=0 Oct 09 10:23:07 crc kubenswrapper[4923]: I1009 10:23:07.864966 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" event={"ID":"75a9acc9-8a59-4a57-abad-bed6920dcc7c","Type":"ContainerDied","Data":"b28eb009542785a30bfc4c524f2bd23830d8e5811c093f37f918fd4660bf9ce9"} Oct 09 10:23:08 crc kubenswrapper[4923]: I1009 10:23:08.884040 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbfdd98cc-7gf4c" event={"ID":"ac145f4f-e2db-4f2f-8842-7256514a4ee2","Type":"ContainerStarted","Data":"65686260d9c999217babf19c7a40e6811eab66897c8aff12eb2ebb1799d8fb91"} Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.199088 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.297718 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-nb\") pod \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.298235 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-sb\") pod \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.298273 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7rnv\" (UniqueName: \"kubernetes.io/projected/75a9acc9-8a59-4a57-abad-bed6920dcc7c-kube-api-access-h7rnv\") pod \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.298947 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-dns-svc\") pod \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.299136 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-config\") pod \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\" (UID: \"75a9acc9-8a59-4a57-abad-bed6920dcc7c\") " Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.311225 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75a9acc9-8a59-4a57-abad-bed6920dcc7c-kube-api-access-h7rnv" (OuterVolumeSpecName: "kube-api-access-h7rnv") pod "75a9acc9-8a59-4a57-abad-bed6920dcc7c" (UID: "75a9acc9-8a59-4a57-abad-bed6920dcc7c"). InnerVolumeSpecName "kube-api-access-h7rnv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.402004 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7rnv\" (UniqueName: \"kubernetes.io/projected/75a9acc9-8a59-4a57-abad-bed6920dcc7c-kube-api-access-h7rnv\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.526427 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-config" (OuterVolumeSpecName: "config") pod "75a9acc9-8a59-4a57-abad-bed6920dcc7c" (UID: "75a9acc9-8a59-4a57-abad-bed6920dcc7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.527876 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "75a9acc9-8a59-4a57-abad-bed6920dcc7c" (UID: "75a9acc9-8a59-4a57-abad-bed6920dcc7c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.541208 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "75a9acc9-8a59-4a57-abad-bed6920dcc7c" (UID: "75a9acc9-8a59-4a57-abad-bed6920dcc7c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.559004 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "75a9acc9-8a59-4a57-abad-bed6920dcc7c" (UID: "75a9acc9-8a59-4a57-abad-bed6920dcc7c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.606143 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.606190 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.606202 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.606217 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/75a9acc9-8a59-4a57-abad-bed6920dcc7c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.921494 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5fbfdd98cc-7gf4c" event={"ID":"ac145f4f-e2db-4f2f-8842-7256514a4ee2","Type":"ContainerStarted","Data":"9175d5ba4887ff06a3705d1704d3b35afed036231c6e71beb4a557131adccf48"} Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.922222 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.935673 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerStarted","Data":"5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f"} Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.952732 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" event={"ID":"75a9acc9-8a59-4a57-abad-bed6920dcc7c","Type":"ContainerDied","Data":"33a2174705feb44a11cab4afd33505168f0ee05745bca9c8c4f104b6d5facba5"} Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.952823 4923 scope.go:117] "RemoveContainer" containerID="b28eb009542785a30bfc4c524f2bd23830d8e5811c093f37f918fd4660bf9ce9" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.953004 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.969633 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5fbfdd98cc-7gf4c" podStartSLOduration=11.969610128 podStartE2EDuration="11.969610128s" podCreationTimestamp="2025-10-09 10:22:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:09.965330219 +0000 UTC m=+1076.033511985" watchObservedRunningTime="2025-10-09 10:23:09.969610128 +0000 UTC m=+1076.037791884" Oct 09 10:23:09 crc kubenswrapper[4923]: I1009 10:23:09.979170 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xq58z" event={"ID":"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07","Type":"ContainerStarted","Data":"23d8fa1b188a24014db4272f7656d4f568788e20dcdef3d0e243e575bb705d49"} Oct 09 10:23:10 crc kubenswrapper[4923]: I1009 10:23:10.002287 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-xq58z" podStartSLOduration=2.8389997190000003 podStartE2EDuration="47.002260062s" podCreationTimestamp="2025-10-09 10:22:23 +0000 UTC" firstStartedPulling="2025-10-09 10:22:24.950039399 +0000 UTC m=+1031.018221155" lastFinishedPulling="2025-10-09 10:23:09.113299742 +0000 UTC m=+1075.181481498" observedRunningTime="2025-10-09 10:23:09.999415913 +0000 UTC m=+1076.067597679" watchObservedRunningTime="2025-10-09 10:23:10.002260062 +0000 UTC m=+1076.070441818" Oct 09 10:23:10 crc kubenswrapper[4923]: I1009 10:23:10.031268 4923 scope.go:117] "RemoveContainer" containerID="3731ed5bf853925dcee54aa6a455b4f1aadc8d102bd3eb9cea29bb1dc1f1a975" Oct 09 10:23:10 crc kubenswrapper[4923]: I1009 10:23:10.070916 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-kgl2b"] Oct 09 10:23:10 crc kubenswrapper[4923]: I1009 10:23:10.093452 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b99bccc6c-kgl2b"] Oct 09 10:23:10 crc kubenswrapper[4923]: E1009 10:23:10.157906 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod75a9acc9_8a59_4a57_abad_bed6920dcc7c.slice/crio-33a2174705feb44a11cab4afd33505168f0ee05745bca9c8c4f104b6d5facba5\": RecentStats: unable to find data in memory cache]" Oct 09 10:23:10 crc kubenswrapper[4923]: I1009 10:23:10.612642 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" path="/var/lib/kubelet/pods/75a9acc9-8a59-4a57-abad-bed6920dcc7c/volumes" Oct 09 10:23:10 crc kubenswrapper[4923]: I1009 10:23:10.999042 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5ll8s" event={"ID":"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd","Type":"ContainerStarted","Data":"e3a247965a5127a1b0ba26eacf32079b5bde1889f0df43458415b5fbe19da36e"} Oct 09 10:23:11 crc kubenswrapper[4923]: I1009 10:23:11.016718 4923 generic.go:334] "Generic (PLEG): container finished" podID="85c95222-efd5-4bd6-ab01-c51eb21abb4c" containerID="379ad10c6f26658939e5b9c00ad61fcc48f49cba5639a288227a541078da67cf" exitCode=0 Oct 09 10:23:11 crc kubenswrapper[4923]: I1009 10:23:11.016866 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2m7bb" event={"ID":"85c95222-efd5-4bd6-ab01-c51eb21abb4c","Type":"ContainerDied","Data":"379ad10c6f26658939e5b9c00ad61fcc48f49cba5639a288227a541078da67cf"} Oct 09 10:23:11 crc kubenswrapper[4923]: I1009 10:23:11.022672 4923 generic.go:334] "Generic (PLEG): container finished" podID="f250b2a7-fb69-4c40-b186-238f74472a44" containerID="8917a51a7c0f6b85a9257086102be6807938116cc331605350a66bcfa7d2e5a0" exitCode=0 Oct 09 10:23:11 crc kubenswrapper[4923]: I1009 10:23:11.024004 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-665vn" event={"ID":"f250b2a7-fb69-4c40-b186-238f74472a44","Type":"ContainerDied","Data":"8917a51a7c0f6b85a9257086102be6807938116cc331605350a66bcfa7d2e5a0"} Oct 09 10:23:11 crc kubenswrapper[4923]: I1009 10:23:11.060672 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-5ll8s" podStartSLOduration=3.47521207 podStartE2EDuration="48.060654864s" podCreationTimestamp="2025-10-09 10:22:23 +0000 UTC" firstStartedPulling="2025-10-09 10:22:24.592774615 +0000 UTC m=+1030.660956371" lastFinishedPulling="2025-10-09 10:23:09.178217409 +0000 UTC m=+1075.246399165" observedRunningTime="2025-10-09 10:23:11.035206389 +0000 UTC m=+1077.103388145" watchObservedRunningTime="2025-10-09 10:23:11.060654864 +0000 UTC m=+1077.128836620" Oct 09 10:23:12 crc kubenswrapper[4923]: I1009 10:23:12.078854 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-855475c586-p22m4" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 09 10:23:12 crc kubenswrapper[4923]: I1009 10:23:12.251345 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5dd576c454-s7hpl" podUID="4d75434e-89d3-464b-af06-fb0ac9f92fd8" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.142:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.142:8443: connect: connection refused" Oct 09 10:23:14 crc kubenswrapper[4923]: I1009 10:23:14.119043 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7b99bccc6c-kgl2b" podUID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.138:5353: i/o timeout" Oct 09 10:23:15 crc kubenswrapper[4923]: I1009 10:23:15.961334 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:23:15 crc kubenswrapper[4923]: I1009 10:23:15.980748 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-665vn" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.065906 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-scripts\") pod \"f250b2a7-fb69-4c40-b186-238f74472a44\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.066299 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-scripts\") pod \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.066559 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-combined-ca-bundle\") pod \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.067119 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-fernet-keys\") pod \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.067278 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-config-data\") pod \"f250b2a7-fb69-4c40-b186-238f74472a44\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.067580 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-config-data\") pod \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.067689 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cvp2r\" (UniqueName: \"kubernetes.io/projected/85c95222-efd5-4bd6-ab01-c51eb21abb4c-kube-api-access-cvp2r\") pod \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.068869 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-combined-ca-bundle\") pod \"f250b2a7-fb69-4c40-b186-238f74472a44\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.068986 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-credential-keys\") pod \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\" (UID: \"85c95222-efd5-4bd6-ab01-c51eb21abb4c\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.069059 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqq2d\" (UniqueName: \"kubernetes.io/projected/f250b2a7-fb69-4c40-b186-238f74472a44-kube-api-access-fqq2d\") pod \"f250b2a7-fb69-4c40-b186-238f74472a44\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.069175 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f250b2a7-fb69-4c40-b186-238f74472a44-logs\") pod \"f250b2a7-fb69-4c40-b186-238f74472a44\" (UID: \"f250b2a7-fb69-4c40-b186-238f74472a44\") " Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.071179 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-scripts" (OuterVolumeSpecName: "scripts") pod "f250b2a7-fb69-4c40-b186-238f74472a44" (UID: "f250b2a7-fb69-4c40-b186-238f74472a44"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.071948 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "85c95222-efd5-4bd6-ab01-c51eb21abb4c" (UID: "85c95222-efd5-4bd6-ab01-c51eb21abb4c"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.074076 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f250b2a7-fb69-4c40-b186-238f74472a44-logs" (OuterVolumeSpecName: "logs") pod "f250b2a7-fb69-4c40-b186-238f74472a44" (UID: "f250b2a7-fb69-4c40-b186-238f74472a44"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.077939 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-scripts" (OuterVolumeSpecName: "scripts") pod "85c95222-efd5-4bd6-ab01-c51eb21abb4c" (UID: "85c95222-efd5-4bd6-ab01-c51eb21abb4c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.078173 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f250b2a7-fb69-4c40-b186-238f74472a44-kube-api-access-fqq2d" (OuterVolumeSpecName: "kube-api-access-fqq2d") pod "f250b2a7-fb69-4c40-b186-238f74472a44" (UID: "f250b2a7-fb69-4c40-b186-238f74472a44"). InnerVolumeSpecName "kube-api-access-fqq2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.081733 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2m7bb" event={"ID":"85c95222-efd5-4bd6-ab01-c51eb21abb4c","Type":"ContainerDied","Data":"b060555b0665e34cfc8d1a492ad44bbf786abc91a225bf2d3b6ca1e8f38a400a"} Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.081796 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b060555b0665e34cfc8d1a492ad44bbf786abc91a225bf2d3b6ca1e8f38a400a" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.081910 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2m7bb" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.085256 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "85c95222-efd5-4bd6-ab01-c51eb21abb4c" (UID: "85c95222-efd5-4bd6-ab01-c51eb21abb4c"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.091555 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-665vn" event={"ID":"f250b2a7-fb69-4c40-b186-238f74472a44","Type":"ContainerDied","Data":"3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a"} Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.091623 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.091741 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-665vn" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.098828 4923 generic.go:334] "Generic (PLEG): container finished" podID="31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" containerID="23d8fa1b188a24014db4272f7656d4f568788e20dcdef3d0e243e575bb705d49" exitCode=0 Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.098931 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xq58z" event={"ID":"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07","Type":"ContainerDied","Data":"23d8fa1b188a24014db4272f7656d4f568788e20dcdef3d0e243e575bb705d49"} Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.111392 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c95222-efd5-4bd6-ab01-c51eb21abb4c-kube-api-access-cvp2r" (OuterVolumeSpecName: "kube-api-access-cvp2r") pod "85c95222-efd5-4bd6-ab01-c51eb21abb4c" (UID: "85c95222-efd5-4bd6-ab01-c51eb21abb4c"). InnerVolumeSpecName "kube-api-access-cvp2r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.172202 4923 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.172239 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqq2d\" (UniqueName: \"kubernetes.io/projected/f250b2a7-fb69-4c40-b186-238f74472a44-kube-api-access-fqq2d\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.172254 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f250b2a7-fb69-4c40-b186-238f74472a44-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.172265 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.172273 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.172283 4923 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.172292 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cvp2r\" (UniqueName: \"kubernetes.io/projected/85c95222-efd5-4bd6-ab01-c51eb21abb4c-kube-api-access-cvp2r\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.173507 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85c95222-efd5-4bd6-ab01-c51eb21abb4c" (UID: "85c95222-efd5-4bd6-ab01-c51eb21abb4c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.192710 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-config-data" (OuterVolumeSpecName: "config-data") pod "f250b2a7-fb69-4c40-b186-238f74472a44" (UID: "f250b2a7-fb69-4c40-b186-238f74472a44"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.198280 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f250b2a7-fb69-4c40-b186-238f74472a44" (UID: "f250b2a7-fb69-4c40-b186-238f74472a44"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.204638 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-config-data" (OuterVolumeSpecName: "config-data") pod "85c95222-efd5-4bd6-ab01-c51eb21abb4c" (UID: "85c95222-efd5-4bd6-ab01-c51eb21abb4c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.273817 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.273870 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.273891 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85c95222-efd5-4bd6-ab01-c51eb21abb4c-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:16 crc kubenswrapper[4923]: I1009 10:23:16.273905 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f250b2a7-fb69-4c40-b186-238f74472a44-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.126218 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerStarted","Data":"4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad"} Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.127148 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5c7cc6d588-zpbfk"] Oct 09 10:23:17 crc kubenswrapper[4923]: E1009 10:23:17.127549 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f250b2a7-fb69-4c40-b186-238f74472a44" containerName="placement-db-sync" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.127566 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f250b2a7-fb69-4c40-b186-238f74472a44" containerName="placement-db-sync" Oct 09 10:23:17 crc kubenswrapper[4923]: E1009 10:23:17.127578 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" containerName="dnsmasq-dns" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.127584 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" containerName="dnsmasq-dns" Oct 09 10:23:17 crc kubenswrapper[4923]: E1009 10:23:17.127612 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" containerName="init" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.127618 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" containerName="init" Oct 09 10:23:17 crc kubenswrapper[4923]: E1009 10:23:17.127645 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c95222-efd5-4bd6-ab01-c51eb21abb4c" containerName="keystone-bootstrap" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.127651 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c95222-efd5-4bd6-ab01-c51eb21abb4c" containerName="keystone-bootstrap" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.127822 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f250b2a7-fb69-4c40-b186-238f74472a44" containerName="placement-db-sync" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.127861 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="75a9acc9-8a59-4a57-abad-bed6920dcc7c" containerName="dnsmasq-dns" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.127873 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c95222-efd5-4bd6-ab01-c51eb21abb4c" containerName="keystone-bootstrap" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.128512 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.133185 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.133357 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.133622 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.133782 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-58pgm" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.133905 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.134085 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.191304 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c7cc6d588-zpbfk"] Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.202005 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-scripts\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.202079 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-fernet-keys\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.202275 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-internal-tls-certs\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.202341 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-config-data\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.202382 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xl9t\" (UniqueName: \"kubernetes.io/projected/4a4c4530-4816-41ac-bdfc-945bb60600c7-kube-api-access-4xl9t\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.202541 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-public-tls-certs\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.202646 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-combined-ca-bundle\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.202798 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-credential-keys\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.312022 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-internal-tls-certs\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.314645 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-config-data\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.314817 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xl9t\" (UniqueName: \"kubernetes.io/projected/4a4c4530-4816-41ac-bdfc-945bb60600c7-kube-api-access-4xl9t\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.315033 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-public-tls-certs\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.315173 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-combined-ca-bundle\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.315344 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-credential-keys\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.315439 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-scripts\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.315472 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-fernet-keys\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.325373 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-84bc56c5fd-mzbzg"] Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.331660 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-credential-keys\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.331736 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-fernet-keys\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.336715 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-scripts\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.337402 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-public-tls-certs\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.338361 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.348392 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-internal-tls-certs\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.351178 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-58dhc" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.351359 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-84bc56c5fd-mzbzg"] Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.353411 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-combined-ca-bundle\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.351972 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.352046 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.352106 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.353526 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.355885 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a4c4530-4816-41ac-bdfc-945bb60600c7-config-data\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.377188 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xl9t\" (UniqueName: \"kubernetes.io/projected/4a4c4530-4816-41ac-bdfc-945bb60600c7-kube-api-access-4xl9t\") pod \"keystone-5c7cc6d588-zpbfk\" (UID: \"4a4c4530-4816-41ac-bdfc-945bb60600c7\") " pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.422416 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-combined-ca-bundle\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.422621 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-scripts\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.422742 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-config-data\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.422944 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-public-tls-certs\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.423056 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8c099f6-0c20-40d9-a47e-35150638b521-logs\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.423149 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptl96\" (UniqueName: \"kubernetes.io/projected/c8c099f6-0c20-40d9-a47e-35150638b521-kube-api-access-ptl96\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.423266 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-internal-tls-certs\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.467468 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.526909 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-internal-tls-certs\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.527005 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-combined-ca-bundle\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.527044 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-scripts\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.527095 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-config-data\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.527117 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-public-tls-certs\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.527159 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8c099f6-0c20-40d9-a47e-35150638b521-logs\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.527191 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptl96\" (UniqueName: \"kubernetes.io/projected/c8c099f6-0c20-40d9-a47e-35150638b521-kube-api-access-ptl96\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.534650 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c8c099f6-0c20-40d9-a47e-35150638b521-logs\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.551923 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-internal-tls-certs\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.551944 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-config-data\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.555549 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-public-tls-certs\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.558346 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptl96\" (UniqueName: \"kubernetes.io/projected/c8c099f6-0c20-40d9-a47e-35150638b521-kube-api-access-ptl96\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.558371 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-combined-ca-bundle\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.558961 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c8c099f6-0c20-40d9-a47e-35150638b521-scripts\") pod \"placement-84bc56c5fd-mzbzg\" (UID: \"c8c099f6-0c20-40d9-a47e-35150638b521\") " pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.607184 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xq58z" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.730642 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-db-sync-config-data\") pod \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.731100 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqb5c\" (UniqueName: \"kubernetes.io/projected/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-kube-api-access-bqb5c\") pod \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.734693 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-combined-ca-bundle\") pod \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\" (UID: \"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07\") " Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.737615 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" (UID: "31f10461-abd0-4c4d-bdb5-3da1f1e6ed07"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.739037 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-kube-api-access-bqb5c" (OuterVolumeSpecName: "kube-api-access-bqb5c") pod "31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" (UID: "31f10461-abd0-4c4d-bdb5-3da1f1e6ed07"). InnerVolumeSpecName "kube-api-access-bqb5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.804170 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.813926 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" (UID: "31f10461-abd0-4c4d-bdb5-3da1f1e6ed07"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.839368 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.839412 4923 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:17 crc kubenswrapper[4923]: I1009 10:23:17.839426 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqb5c\" (UniqueName: \"kubernetes.io/projected/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07-kube-api-access-bqb5c\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.044231 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c7cc6d588-zpbfk"] Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.149956 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c7cc6d588-zpbfk" event={"ID":"4a4c4530-4816-41ac-bdfc-945bb60600c7","Type":"ContainerStarted","Data":"865ecc9c9d49da97e7510d9dc8b1fd3c534015720a67c7a99cf244f9f47ff736"} Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.170893 4923 generic.go:334] "Generic (PLEG): container finished" podID="0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" containerID="e3a247965a5127a1b0ba26eacf32079b5bde1889f0df43458415b5fbe19da36e" exitCode=0 Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.171002 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5ll8s" event={"ID":"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd","Type":"ContainerDied","Data":"e3a247965a5127a1b0ba26eacf32079b5bde1889f0df43458415b5fbe19da36e"} Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.176266 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-xq58z" event={"ID":"31f10461-abd0-4c4d-bdb5-3da1f1e6ed07","Type":"ContainerDied","Data":"d8ad35af505b36c417b15e9377d7dad77649d92bc90d5ee0bbf05f6824e50748"} Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.176350 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8ad35af505b36c417b15e9377d7dad77649d92bc90d5ee0bbf05f6824e50748" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.176350 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-xq58z" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.339353 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-84bc56c5fd-mzbzg"] Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.392354 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-55d7d7d987-nr6w5"] Oct 09 10:23:18 crc kubenswrapper[4923]: E1009 10:23:18.392919 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" containerName="barbican-db-sync" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.392940 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" containerName="barbican-db-sync" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.393120 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" containerName="barbican-db-sync" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.394298 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.401395 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-f825x" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.402020 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.402248 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.419905 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7bc74dbf64-gj94f"] Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.425641 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.430516 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.460660 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-config-data\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.460724 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-combined-ca-bundle\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.460779 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8334ffed-82c7-483f-a947-6f3c132965f8-combined-ca-bundle\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.460836 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8334ffed-82c7-483f-a947-6f3c132965f8-logs\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.460859 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8334ffed-82c7-483f-a947-6f3c132965f8-config-data-custom\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.460887 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-logs\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.460934 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8334ffed-82c7-483f-a947-6f3c132965f8-config-data\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.460983 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6xtw\" (UniqueName: \"kubernetes.io/projected/8334ffed-82c7-483f-a947-6f3c132965f8-kube-api-access-d6xtw\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.461024 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-config-data-custom\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.461065 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kljqz\" (UniqueName: \"kubernetes.io/projected/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-kube-api-access-kljqz\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.467347 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-55d7d7d987-nr6w5"] Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.527278 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bc74dbf64-gj94f"] Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.563227 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8334ffed-82c7-483f-a947-6f3c132965f8-config-data\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.563506 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6xtw\" (UniqueName: \"kubernetes.io/projected/8334ffed-82c7-483f-a947-6f3c132965f8-kube-api-access-d6xtw\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.563611 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-config-data-custom\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.563920 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kljqz\" (UniqueName: \"kubernetes.io/projected/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-kube-api-access-kljqz\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.564024 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-config-data\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.564170 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-combined-ca-bundle\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.564271 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8334ffed-82c7-483f-a947-6f3c132965f8-combined-ca-bundle\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.564357 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8334ffed-82c7-483f-a947-6f3c132965f8-logs\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.564430 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8334ffed-82c7-483f-a947-6f3c132965f8-config-data-custom\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.564520 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-logs\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.565055 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-logs\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.567664 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8334ffed-82c7-483f-a947-6f3c132965f8-logs\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.570695 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8334ffed-82c7-483f-a947-6f3c132965f8-config-data\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.575784 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8334ffed-82c7-483f-a947-6f3c132965f8-config-data-custom\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.576391 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8334ffed-82c7-483f-a947-6f3c132965f8-combined-ca-bundle\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.578672 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-combined-ca-bundle\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.583185 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-config-data\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.583644 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-config-data-custom\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.595540 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6xtw\" (UniqueName: \"kubernetes.io/projected/8334ffed-82c7-483f-a947-6f3c132965f8-kube-api-access-d6xtw\") pod \"barbican-keystone-listener-7bc74dbf64-gj94f\" (UID: \"8334ffed-82c7-483f-a947-6f3c132965f8\") " pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.601421 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kljqz\" (UniqueName: \"kubernetes.io/projected/169c7b93-c284-4cf2-bc5c-c5f3a7c7277d-kube-api-access-kljqz\") pod \"barbican-worker-55d7d7d987-nr6w5\" (UID: \"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d\") " pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.649913 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mqml4"] Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.652061 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mqml4"] Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.652172 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.664308 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6744bb4d8-9hkdw"] Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.666006 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.678960 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.680139 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6744bb4d8-9hkdw"] Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.753906 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-55d7d7d987-nr6w5" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.769541 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4d125b1-feb1-43f8-94ad-417f4e163822-logs\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.769650 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjzzt\" (UniqueName: \"kubernetes.io/projected/e4d125b1-feb1-43f8-94ad-417f4e163822-kube-api-access-zjzzt\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.770042 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-combined-ca-bundle\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.770114 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdcb2\" (UniqueName: \"kubernetes.io/projected/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-kube-api-access-wdcb2\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.770263 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data-custom\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.770304 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-dns-svc\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.770382 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.770447 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.770499 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-config\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.770528 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.809719 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.876225 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-combined-ca-bundle\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.876293 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdcb2\" (UniqueName: \"kubernetes.io/projected/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-kube-api-access-wdcb2\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.877805 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data-custom\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.877836 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-dns-svc\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.877874 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.877903 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.877947 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-config\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.877970 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.878044 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4d125b1-feb1-43f8-94ad-417f4e163822-logs\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.878101 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjzzt\" (UniqueName: \"kubernetes.io/projected/e4d125b1-feb1-43f8-94ad-417f4e163822-kube-api-access-zjzzt\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.880942 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-sb\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.881782 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-dns-svc\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.882144 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4d125b1-feb1-43f8-94ad-417f4e163822-logs\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.883721 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-combined-ca-bundle\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.885537 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-config\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.895869 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-nb\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.896481 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.898662 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data-custom\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.904150 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjzzt\" (UniqueName: \"kubernetes.io/projected/e4d125b1-feb1-43f8-94ad-417f4e163822-kube-api-access-zjzzt\") pod \"barbican-api-6744bb4d8-9hkdw\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:18 crc kubenswrapper[4923]: I1009 10:23:18.908456 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdcb2\" (UniqueName: \"kubernetes.io/projected/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-kube-api-access-wdcb2\") pod \"dnsmasq-dns-6dc8d75dbf-mqml4\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.019445 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.030644 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.222399 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c7cc6d588-zpbfk" event={"ID":"4a4c4530-4816-41ac-bdfc-945bb60600c7","Type":"ContainerStarted","Data":"c91758719af009d7b9a82d8de86c53d18cacf537c5ed06d6f09289f14df0ae7e"} Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.222851 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.250838 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84bc56c5fd-mzbzg" event={"ID":"c8c099f6-0c20-40d9-a47e-35150638b521","Type":"ContainerStarted","Data":"cf878cf2f543187c55d36f294f5feaea04c870738840945c511616257265bd9b"} Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.468410 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5c7cc6d588-zpbfk" podStartSLOduration=2.468380828 podStartE2EDuration="2.468380828s" podCreationTimestamp="2025-10-09 10:23:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:19.258937488 +0000 UTC m=+1085.327119244" watchObservedRunningTime="2025-10-09 10:23:19.468380828 +0000 UTC m=+1085.536562604" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.472879 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-55d7d7d987-nr6w5"] Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.568409 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7bc74dbf64-gj94f"] Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.722180 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.773529 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mqml4"] Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.814013 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-scripts\") pod \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.814161 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-combined-ca-bundle\") pod \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.814223 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7rgc\" (UniqueName: \"kubernetes.io/projected/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-kube-api-access-r7rgc\") pod \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.814333 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-db-sync-config-data\") pod \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.814437 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-etc-machine-id\") pod \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.814535 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-config-data\") pod \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\" (UID: \"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd\") " Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.816956 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" (UID: "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.822332 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" (UID: "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.822615 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-scripts" (OuterVolumeSpecName: "scripts") pod "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" (UID: "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.828271 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-kube-api-access-r7rgc" (OuterVolumeSpecName: "kube-api-access-r7rgc") pod "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" (UID: "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd"). InnerVolumeSpecName "kube-api-access-r7rgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.868272 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" (UID: "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.907918 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-config-data" (OuterVolumeSpecName: "config-data") pod "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" (UID: "0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.918498 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.918536 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.918546 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.918560 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7rgc\" (UniqueName: \"kubernetes.io/projected/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-kube-api-access-r7rgc\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.918571 4923 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.918583 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:19 crc kubenswrapper[4923]: I1009 10:23:19.965967 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6744bb4d8-9hkdw"] Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.264154 4923 generic.go:334] "Generic (PLEG): container finished" podID="f6cf5e15-cff3-4f80-92a8-af55f699eaa1" containerID="479310845d56c2530f5b714401108d6fa278310de099f41ee227d969f2fd7182" exitCode=0 Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.264262 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" event={"ID":"f6cf5e15-cff3-4f80-92a8-af55f699eaa1","Type":"ContainerDied","Data":"479310845d56c2530f5b714401108d6fa278310de099f41ee227d969f2fd7182"} Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.264330 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" event={"ID":"f6cf5e15-cff3-4f80-92a8-af55f699eaa1","Type":"ContainerStarted","Data":"aee24e31fa83043a8df2443df9301f536b5e597bfeb432072e60b307781c6f2c"} Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.269342 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-5ll8s" event={"ID":"0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd","Type":"ContainerDied","Data":"2dfd5b2fdc6e79f2067348aad875a022677b2033adfa4be051c6ac4337c16f56"} Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.269380 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-5ll8s" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.269403 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dfd5b2fdc6e79f2067348aad875a022677b2033adfa4be051c6ac4337c16f56" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.276117 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84bc56c5fd-mzbzg" event={"ID":"c8c099f6-0c20-40d9-a47e-35150638b521","Type":"ContainerStarted","Data":"df8ea114d5b362e389982582c4cb12a7604cee39c36573e57732018252c9e259"} Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.277696 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" event={"ID":"8334ffed-82c7-483f-a947-6f3c132965f8","Type":"ContainerStarted","Data":"455d5048e9f67fb96750e894681f925d5f0739556ea3895442db9881e5e750b6"} Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.291175 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6744bb4d8-9hkdw" event={"ID":"e4d125b1-feb1-43f8-94ad-417f4e163822","Type":"ContainerStarted","Data":"715b146c872f0b5a9d2f682807657969f373dfb5f2015d1b418b54e3b7d9bff3"} Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.305142 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-55d7d7d987-nr6w5" event={"ID":"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d","Type":"ContainerStarted","Data":"cf548f5e2f10658dca46a769f5678694ba4a0d56b2f0f745d74ffd8023d83e9f"} Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.584027 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mqml4"] Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.643931 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 10:23:20 crc kubenswrapper[4923]: E1009 10:23:20.644301 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" containerName="cinder-db-sync" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.644324 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" containerName="cinder-db-sync" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.644548 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" containerName="cinder-db-sync" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.653096 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.657866 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.669431 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.669483 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kzbkw" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.669830 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.670129 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.688996 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85494b87f-hlckn"] Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.691039 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.703820 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-hlckn"] Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.758617 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.762681 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-dns-svc\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.762823 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72e46628-da39-4445-a4e2-0d372fce2716-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.762880 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-sb\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.762977 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.763311 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-scripts\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.763437 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgvw8\" (UniqueName: \"kubernetes.io/projected/72e46628-da39-4445-a4e2-0d372fce2716-kube-api-access-hgvw8\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.763600 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-nb\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.763710 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.763783 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-config\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.764036 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2wl5\" (UniqueName: \"kubernetes.io/projected/51933409-9af8-4cd7-8642-6d4f675068d7-kube-api-access-b2wl5\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.798430 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.817326 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.824922 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.835743 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.867217 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.867566 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.867651 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-config\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.867737 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29f311b-49a9-4251-b640-79dea462412f-logs\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.867857 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data-custom\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.867941 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkk8f\" (UniqueName: \"kubernetes.io/projected/d29f311b-49a9-4251-b640-79dea462412f-kube-api-access-kkk8f\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868035 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2wl5\" (UniqueName: \"kubernetes.io/projected/51933409-9af8-4cd7-8642-6d4f675068d7-kube-api-access-b2wl5\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868147 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868220 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-dns-svc\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868310 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72e46628-da39-4445-a4e2-0d372fce2716-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868387 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-sb\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868492 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868581 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868656 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-scripts\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868741 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-scripts\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868874 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgvw8\" (UniqueName: \"kubernetes.io/projected/72e46628-da39-4445-a4e2-0d372fce2716-kube-api-access-hgvw8\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.868960 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d29f311b-49a9-4251-b640-79dea462412f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.869036 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-nb\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.870144 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-nb\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.871038 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72e46628-da39-4445-a4e2-0d372fce2716-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.871876 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-config\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.884668 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-dns-svc\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.889138 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-sb\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.895693 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.900258 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-scripts\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.909942 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2wl5\" (UniqueName: \"kubernetes.io/projected/51933409-9af8-4cd7-8642-6d4f675068d7-kube-api-access-b2wl5\") pod \"dnsmasq-dns-85494b87f-hlckn\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.921093 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.943087 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.945048 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgvw8\" (UniqueName: \"kubernetes.io/projected/72e46628-da39-4445-a4e2-0d372fce2716-kube-api-access-hgvw8\") pod \"cinder-scheduler-0\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.970766 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d29f311b-49a9-4251-b640-79dea462412f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.970842 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.970887 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29f311b-49a9-4251-b640-79dea462412f-logs\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.970916 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data-custom\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.970922 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d29f311b-49a9-4251-b640-79dea462412f-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.970942 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkk8f\" (UniqueName: \"kubernetes.io/projected/d29f311b-49a9-4251-b640-79dea462412f-kube-api-access-kkk8f\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.971370 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.971425 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-scripts\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.972892 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29f311b-49a9-4251-b640-79dea462412f-logs\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.980496 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.981191 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data-custom\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.982373 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-scripts\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.982809 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:20 crc kubenswrapper[4923]: I1009 10:23:20.995179 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkk8f\" (UniqueName: \"kubernetes.io/projected/d29f311b-49a9-4251-b640-79dea462412f-kube-api-access-kkk8f\") pod \"cinder-api-0\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " pod="openstack/cinder-api-0" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.006654 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.094329 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.121519 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.371009 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-84bc56c5fd-mzbzg" event={"ID":"c8c099f6-0c20-40d9-a47e-35150638b521","Type":"ContainerStarted","Data":"2503476f80668184f6d2c5a0cbbf91962e7309fb3186f8c6266b307daaf2c7cb"} Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.371552 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.371580 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.406425 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-84bc56c5fd-mzbzg" podStartSLOduration=4.406400823 podStartE2EDuration="4.406400823s" podCreationTimestamp="2025-10-09 10:23:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:21.403900544 +0000 UTC m=+1087.472082300" watchObservedRunningTime="2025-10-09 10:23:21.406400823 +0000 UTC m=+1087.474582579" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.418688 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6744bb4d8-9hkdw" event={"ID":"e4d125b1-feb1-43f8-94ad-417f4e163822","Type":"ContainerStarted","Data":"7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad"} Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.418750 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6744bb4d8-9hkdw" event={"ID":"e4d125b1-feb1-43f8-94ad-417f4e163822","Type":"ContainerStarted","Data":"0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8"} Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.418995 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.437513 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" event={"ID":"f6cf5e15-cff3-4f80-92a8-af55f699eaa1","Type":"ContainerStarted","Data":"94e5a7f36cc4bbe44e9b4666b3f5b68a9fe8fc21d7bd5152e9ee6b42c7fc6547"} Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.437957 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" podUID="f6cf5e15-cff3-4f80-92a8-af55f699eaa1" containerName="dnsmasq-dns" containerID="cri-o://94e5a7f36cc4bbe44e9b4666b3f5b68a9fe8fc21d7bd5152e9ee6b42c7fc6547" gracePeriod=10 Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.451900 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.469879 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6744bb4d8-9hkdw" podStartSLOduration=3.46984338 podStartE2EDuration="3.46984338s" podCreationTimestamp="2025-10-09 10:23:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:21.451429941 +0000 UTC m=+1087.519611707" watchObservedRunningTime="2025-10-09 10:23:21.46984338 +0000 UTC m=+1087.538025146" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.501895 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" podStartSLOduration=3.501866628 podStartE2EDuration="3.501866628s" podCreationTimestamp="2025-10-09 10:23:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:21.499091491 +0000 UTC m=+1087.567273247" watchObservedRunningTime="2025-10-09 10:23:21.501866628 +0000 UTC m=+1087.570048384" Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.558779 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 10:23:21 crc kubenswrapper[4923]: I1009 10:23:21.870898 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-hlckn"] Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.037685 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.451053 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-hlckn" event={"ID":"51933409-9af8-4cd7-8642-6d4f675068d7","Type":"ContainerStarted","Data":"136ea258b6bb1ff3fcd8d4d7251d283af686e9f96538eb629a4ab119af745cb9"} Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.453446 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72e46628-da39-4445-a4e2-0d372fce2716","Type":"ContainerStarted","Data":"63586d2cbbcfc7bb7be2346d3733da333bb6f9715c18e6e41a444358bbc103b2"} Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.460889 4923 generic.go:334] "Generic (PLEG): container finished" podID="f6cf5e15-cff3-4f80-92a8-af55f699eaa1" containerID="94e5a7f36cc4bbe44e9b4666b3f5b68a9fe8fc21d7bd5152e9ee6b42c7fc6547" exitCode=0 Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.462275 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" event={"ID":"f6cf5e15-cff3-4f80-92a8-af55f699eaa1","Type":"ContainerDied","Data":"94e5a7f36cc4bbe44e9b4666b3f5b68a9fe8fc21d7bd5152e9ee6b42c7fc6547"} Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.462923 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.753099 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.852682 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-config\") pod \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.852733 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-dns-svc\") pod \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.852813 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-sb\") pod \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.852892 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-nb\") pod \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.852923 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdcb2\" (UniqueName: \"kubernetes.io/projected/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-kube-api-access-wdcb2\") pod \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\" (UID: \"f6cf5e15-cff3-4f80-92a8-af55f699eaa1\") " Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.858608 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-kube-api-access-wdcb2" (OuterVolumeSpecName: "kube-api-access-wdcb2") pod "f6cf5e15-cff3-4f80-92a8-af55f699eaa1" (UID: "f6cf5e15-cff3-4f80-92a8-af55f699eaa1"). InnerVolumeSpecName "kube-api-access-wdcb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.915351 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f6cf5e15-cff3-4f80-92a8-af55f699eaa1" (UID: "f6cf5e15-cff3-4f80-92a8-af55f699eaa1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.940149 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f6cf5e15-cff3-4f80-92a8-af55f699eaa1" (UID: "f6cf5e15-cff3-4f80-92a8-af55f699eaa1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.954131 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-config" (OuterVolumeSpecName: "config") pod "f6cf5e15-cff3-4f80-92a8-af55f699eaa1" (UID: "f6cf5e15-cff3-4f80-92a8-af55f699eaa1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.956482 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.956524 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdcb2\" (UniqueName: \"kubernetes.io/projected/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-kube-api-access-wdcb2\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.956537 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.956547 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:22 crc kubenswrapper[4923]: I1009 10:23:22.958598 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f6cf5e15-cff3-4f80-92a8-af55f699eaa1" (UID: "f6cf5e15-cff3-4f80-92a8-af55f699eaa1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:23 crc kubenswrapper[4923]: I1009 10:23:23.057940 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f6cf5e15-cff3-4f80-92a8-af55f699eaa1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:23 crc kubenswrapper[4923]: I1009 10:23:23.495071 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" event={"ID":"f6cf5e15-cff3-4f80-92a8-af55f699eaa1","Type":"ContainerDied","Data":"aee24e31fa83043a8df2443df9301f536b5e597bfeb432072e60b307781c6f2c"} Oct 09 10:23:23 crc kubenswrapper[4923]: I1009 10:23:23.495098 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6dc8d75dbf-mqml4" Oct 09 10:23:23 crc kubenswrapper[4923]: I1009 10:23:23.495775 4923 scope.go:117] "RemoveContainer" containerID="94e5a7f36cc4bbe44e9b4666b3f5b68a9fe8fc21d7bd5152e9ee6b42c7fc6547" Oct 09 10:23:23 crc kubenswrapper[4923]: I1009 10:23:23.497601 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d29f311b-49a9-4251-b640-79dea462412f","Type":"ContainerStarted","Data":"0a4299f657e05a8e0357268cc573bbfc564a45b92474ef463f4be735b4b81403"} Oct 09 10:23:23 crc kubenswrapper[4923]: I1009 10:23:23.548143 4923 scope.go:117] "RemoveContainer" containerID="479310845d56c2530f5b714401108d6fa278310de099f41ee227d969f2fd7182" Oct 09 10:23:23 crc kubenswrapper[4923]: I1009 10:23:23.560746 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mqml4"] Oct 09 10:23:23 crc kubenswrapper[4923]: I1009 10:23:23.570639 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6dc8d75dbf-mqml4"] Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.526196 4923 generic.go:334] "Generic (PLEG): container finished" podID="51933409-9af8-4cd7-8642-6d4f675068d7" containerID="df3599959384db95c3eef973f1475b7365cfd91649b14fc382f41be572fa9e83" exitCode=0 Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.527112 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-hlckn" event={"ID":"51933409-9af8-4cd7-8642-6d4f675068d7","Type":"ContainerDied","Data":"df3599959384db95c3eef973f1475b7365cfd91649b14fc382f41be572fa9e83"} Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.534663 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.542457 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" event={"ID":"8334ffed-82c7-483f-a947-6f3c132965f8","Type":"ContainerStarted","Data":"a7667f856b389b3e5d6212c4f1b0395f6761594c33c6e3b12074f8ad2746aee6"} Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.544734 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-55d7d7d987-nr6w5" event={"ID":"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d","Type":"ContainerStarted","Data":"e3785724f4262da5fd4a4183997108d70d51ee8e12edca81586e7b01721462a4"} Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.560507 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.605266 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.605458 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.619146 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6cf5e15-cff3-4f80-92a8-af55f699eaa1" path="/var/lib/kubelet/pods/f6cf5e15-cff3-4f80-92a8-af55f699eaa1/volumes" Oct 09 10:23:24 crc kubenswrapper[4923]: I1009 10:23:24.750374 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-855475c586-p22m4" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.203034 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-9f849b6cd-m82rj"] Oct 09 10:23:25 crc kubenswrapper[4923]: E1009 10:23:25.203697 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6cf5e15-cff3-4f80-92a8-af55f699eaa1" containerName="init" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.203720 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6cf5e15-cff3-4f80-92a8-af55f699eaa1" containerName="init" Oct 09 10:23:25 crc kubenswrapper[4923]: E1009 10:23:25.203734 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6cf5e15-cff3-4f80-92a8-af55f699eaa1" containerName="dnsmasq-dns" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.203743 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6cf5e15-cff3-4f80-92a8-af55f699eaa1" containerName="dnsmasq-dns" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.204006 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6cf5e15-cff3-4f80-92a8-af55f699eaa1" containerName="dnsmasq-dns" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.205088 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.210995 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.211543 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9stv\" (UniqueName: \"kubernetes.io/projected/45877ffa-4abb-489a-9bad-770cbb5f231a-kube-api-access-c9stv\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.211585 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45877ffa-4abb-489a-9bad-770cbb5f231a-logs\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.211655 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-public-tls-certs\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.211693 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-combined-ca-bundle\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.211721 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-config-data-custom\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.211762 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-config-data\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.211830 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-internal-tls-certs\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.214635 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.235831 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9f849b6cd-m82rj"] Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.314731 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45877ffa-4abb-489a-9bad-770cbb5f231a-logs\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.314973 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-public-tls-certs\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.315072 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-combined-ca-bundle\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.315146 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-config-data-custom\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.315232 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-config-data\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.315347 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/45877ffa-4abb-489a-9bad-770cbb5f231a-logs\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.315561 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-internal-tls-certs\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.315649 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c9stv\" (UniqueName: \"kubernetes.io/projected/45877ffa-4abb-489a-9bad-770cbb5f231a-kube-api-access-c9stv\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.328140 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-combined-ca-bundle\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.338878 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-config-data-custom\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.340523 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-internal-tls-certs\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.351328 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9stv\" (UniqueName: \"kubernetes.io/projected/45877ffa-4abb-489a-9bad-770cbb5f231a-kube-api-access-c9stv\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.362561 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-public-tls-certs\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.377016 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45877ffa-4abb-489a-9bad-770cbb5f231a-config-data\") pod \"barbican-api-9f849b6cd-m82rj\" (UID: \"45877ffa-4abb-489a-9bad-770cbb5f231a\") " pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.558320 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.639330 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d29f311b-49a9-4251-b640-79dea462412f","Type":"ContainerStarted","Data":"1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c"} Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.646484 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72e46628-da39-4445-a4e2-0d372fce2716","Type":"ContainerStarted","Data":"f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47"} Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.651688 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-55d7d7d987-nr6w5" event={"ID":"169c7b93-c284-4cf2-bc5c-c5f3a7c7277d","Type":"ContainerStarted","Data":"69f092e5ed0f79d09b42e4d758681f3e6befabdb0082c219703bf325b08fc685"} Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.687643 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-55d7d7d987-nr6w5" podStartSLOduration=3.723601727 podStartE2EDuration="7.687611913s" podCreationTimestamp="2025-10-09 10:23:18 +0000 UTC" firstStartedPulling="2025-10-09 10:23:19.525650684 +0000 UTC m=+1085.593832440" lastFinishedPulling="2025-10-09 10:23:23.48966087 +0000 UTC m=+1089.557842626" observedRunningTime="2025-10-09 10:23:25.677224355 +0000 UTC m=+1091.745406111" watchObservedRunningTime="2025-10-09 10:23:25.687611913 +0000 UTC m=+1091.755793669" Oct 09 10:23:25 crc kubenswrapper[4923]: I1009 10:23:25.753513 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:23:26 crc kubenswrapper[4923]: W1009 10:23:26.768276 4923 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6cf5e15_cff3_4f80_92a8_af55f699eaa1.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6cf5e15_cff3_4f80_92a8_af55f699eaa1.slice: no such file or directory Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.016718 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-855475c586-p22m4" Oct 09 10:23:27 crc kubenswrapper[4923]: E1009 10:23:27.035728 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c688fdd_7d04_4a0a_9074_94c9f1b8a5fd.slice/crio-2dfd5b2fdc6e79f2067348aad875a022677b2033adfa4be051c6ac4337c16f56\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode39f0227_4155_4d15_9743_d81095c77c0d.slice/crio-conmon-146a9d63b4b576ce6c37a77aa8c2b45c7a68cf6c37f54f1136ced8d817fc95d8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf250b2a7_fb69_4c40_b186_238f74472a44.slice/crio-3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode39f0227_4155_4d15_9743_d81095c77c0d.slice/crio-146a9d63b4b576ce6c37a77aa8c2b45c7a68cf6c37f54f1136ced8d817fc95d8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode39f0227_4155_4d15_9743_d81095c77c0d.slice/crio-conmon-230fb6ba9b99d46c0ba822b55f186e30ac29a85fef94e58740f5ad27d43191a3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0c688fdd_7d04_4a0a_9074_94c9f1b8a5fd.slice\": RecentStats: unable to find data in memory cache]" Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.139288 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5dd576c454-s7hpl" Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.149498 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-6744bb4d8-9hkdw" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.222102 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-855475c586-p22m4"] Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.683452 4923 generic.go:334] "Generic (PLEG): container finished" podID="e39f0227-4155-4d15-9743-d81095c77c0d" containerID="146a9d63b4b576ce6c37a77aa8c2b45c7a68cf6c37f54f1136ced8d817fc95d8" exitCode=137 Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.683490 4923 generic.go:334] "Generic (PLEG): container finished" podID="e39f0227-4155-4d15-9743-d81095c77c0d" containerID="230fb6ba9b99d46c0ba822b55f186e30ac29a85fef94e58740f5ad27d43191a3" exitCode=137 Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.683943 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-855475c586-p22m4" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon-log" containerID="cri-o://fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2" gracePeriod=30 Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.684365 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c85c4b745-jf7ck" event={"ID":"e39f0227-4155-4d15-9743-d81095c77c0d","Type":"ContainerDied","Data":"146a9d63b4b576ce6c37a77aa8c2b45c7a68cf6c37f54f1136ced8d817fc95d8"} Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.684473 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c85c4b745-jf7ck" event={"ID":"e39f0227-4155-4d15-9743-d81095c77c0d","Type":"ContainerDied","Data":"230fb6ba9b99d46c0ba822b55f186e30ac29a85fef94e58740f5ad27d43191a3"} Oct 09 10:23:27 crc kubenswrapper[4923]: I1009 10:23:27.684531 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-855475c586-p22m4" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon" containerID="cri-o://b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667" gracePeriod=30 Oct 09 10:23:28 crc kubenswrapper[4923]: I1009 10:23:28.662949 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5fbfdd98cc-7gf4c" Oct 09 10:23:28 crc kubenswrapper[4923]: I1009 10:23:28.743360 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-665d85ccb-zkkd2"] Oct 09 10:23:28 crc kubenswrapper[4923]: I1009 10:23:28.743856 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-665d85ccb-zkkd2" podUID="f9c24b4f-b830-494a-9789-745e176ffac6" containerName="neutron-api" containerID="cri-o://b95dac3af77883b7de07aef6630527ed09e7cfba4c7092318dd041d2804ba167" gracePeriod=30 Oct 09 10:23:28 crc kubenswrapper[4923]: I1009 10:23:28.744701 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-665d85ccb-zkkd2" podUID="f9c24b4f-b830-494a-9789-745e176ffac6" containerName="neutron-httpd" containerID="cri-o://86bf1e0b381981b4936ffb857b09e0bfb1091b6d8d9c1a65c1150d5f4cfaa492" gracePeriod=30 Oct 09 10:23:29 crc kubenswrapper[4923]: I1009 10:23:29.718200 4923 generic.go:334] "Generic (PLEG): container finished" podID="f9c24b4f-b830-494a-9789-745e176ffac6" containerID="86bf1e0b381981b4936ffb857b09e0bfb1091b6d8d9c1a65c1150d5f4cfaa492" exitCode=0 Oct 09 10:23:29 crc kubenswrapper[4923]: I1009 10:23:29.718644 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-665d85ccb-zkkd2" event={"ID":"f9c24b4f-b830-494a-9789-745e176ffac6","Type":"ContainerDied","Data":"86bf1e0b381981b4936ffb857b09e0bfb1091b6d8d9c1a65c1150d5f4cfaa492"} Oct 09 10:23:30 crc kubenswrapper[4923]: I1009 10:23:30.573023 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:30 crc kubenswrapper[4923]: I1009 10:23:30.573997 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.087821 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.185368 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e39f0227-4155-4d15-9743-d81095c77c0d-horizon-secret-key\") pod \"e39f0227-4155-4d15-9743-d81095c77c0d\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.185675 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx57v\" (UniqueName: \"kubernetes.io/projected/e39f0227-4155-4d15-9743-d81095c77c0d-kube-api-access-gx57v\") pod \"e39f0227-4155-4d15-9743-d81095c77c0d\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.185731 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e39f0227-4155-4d15-9743-d81095c77c0d-logs\") pod \"e39f0227-4155-4d15-9743-d81095c77c0d\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.185857 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-config-data\") pod \"e39f0227-4155-4d15-9743-d81095c77c0d\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.185907 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-scripts\") pod \"e39f0227-4155-4d15-9743-d81095c77c0d\" (UID: \"e39f0227-4155-4d15-9743-d81095c77c0d\") " Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.187007 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e39f0227-4155-4d15-9743-d81095c77c0d-logs" (OuterVolumeSpecName: "logs") pod "e39f0227-4155-4d15-9743-d81095c77c0d" (UID: "e39f0227-4155-4d15-9743-d81095c77c0d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.191992 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e39f0227-4155-4d15-9743-d81095c77c0d-kube-api-access-gx57v" (OuterVolumeSpecName: "kube-api-access-gx57v") pod "e39f0227-4155-4d15-9743-d81095c77c0d" (UID: "e39f0227-4155-4d15-9743-d81095c77c0d"). InnerVolumeSpecName "kube-api-access-gx57v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.195058 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e39f0227-4155-4d15-9743-d81095c77c0d-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "e39f0227-4155-4d15-9743-d81095c77c0d" (UID: "e39f0227-4155-4d15-9743-d81095c77c0d"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.217147 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-scripts" (OuterVolumeSpecName: "scripts") pod "e39f0227-4155-4d15-9743-d81095c77c0d" (UID: "e39f0227-4155-4d15-9743-d81095c77c0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.222927 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-config-data" (OuterVolumeSpecName: "config-data") pod "e39f0227-4155-4d15-9743-d81095c77c0d" (UID: "e39f0227-4155-4d15-9743-d81095c77c0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.289442 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx57v\" (UniqueName: \"kubernetes.io/projected/e39f0227-4155-4d15-9743-d81095c77c0d-kube-api-access-gx57v\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.289496 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e39f0227-4155-4d15-9743-d81095c77c0d-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.292059 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.292079 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e39f0227-4155-4d15-9743-d81095c77c0d-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.292096 4923 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e39f0227-4155-4d15-9743-d81095c77c0d-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.359605 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-9f849b6cd-m82rj"] Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.788077 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72e46628-da39-4445-a4e2-0d372fce2716","Type":"ContainerStarted","Data":"1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9"} Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.795001 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7c85c4b745-jf7ck" event={"ID":"e39f0227-4155-4d15-9743-d81095c77c0d","Type":"ContainerDied","Data":"9905da73f429ef7103df51c18942a7ce0a8d2b29d5462a5c92db7d4c081d2087"} Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.795077 4923 scope.go:117] "RemoveContainer" containerID="146a9d63b4b576ce6c37a77aa8c2b45c7a68cf6c37f54f1136ced8d817fc95d8" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.795518 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7c85c4b745-jf7ck" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.805295 4923 generic.go:334] "Generic (PLEG): container finished" podID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerID="b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667" exitCode=0 Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.805431 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855475c586-p22m4" event={"ID":"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b","Type":"ContainerDied","Data":"b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667"} Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.808458 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9f849b6cd-m82rj" event={"ID":"45877ffa-4abb-489a-9bad-770cbb5f231a","Type":"ContainerStarted","Data":"b69299b1b74f426da3dfee06d07e5ddb63f26ff46f0ae9657dfea12d820705eb"} Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.808556 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9f849b6cd-m82rj" event={"ID":"45877ffa-4abb-489a-9bad-770cbb5f231a","Type":"ContainerStarted","Data":"edd175f67f82efb3f58abf1942110eb6b8bd35379ca94691d425fc661f98a996"} Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.813635 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=9.535777279 podStartE2EDuration="11.813573303s" podCreationTimestamp="2025-10-09 10:23:20 +0000 UTC" firstStartedPulling="2025-10-09 10:23:21.609595521 +0000 UTC m=+1087.677777277" lastFinishedPulling="2025-10-09 10:23:23.887391545 +0000 UTC m=+1089.955573301" observedRunningTime="2025-10-09 10:23:31.809267174 +0000 UTC m=+1097.877448950" watchObservedRunningTime="2025-10-09 10:23:31.813573303 +0000 UTC m=+1097.881755059" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.822006 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerStarted","Data":"13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6"} Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.822247 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="ceilometer-central-agent" containerID="cri-o://b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239" gracePeriod=30 Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.822363 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.822884 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="proxy-httpd" containerID="cri-o://13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6" gracePeriod=30 Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.822983 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="sg-core" containerID="cri-o://4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad" gracePeriod=30 Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.823045 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="ceilometer-notification-agent" containerID="cri-o://5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f" gracePeriod=30 Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.848702 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-hlckn" event={"ID":"51933409-9af8-4cd7-8642-6d4f675068d7","Type":"ContainerStarted","Data":"6a8a0d5cd2b6feda4d71058d60caadf91433abbd2a4b7b65a9cd7f1f63412af1"} Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.850140 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.861766 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" event={"ID":"8334ffed-82c7-483f-a947-6f3c132965f8","Type":"ContainerStarted","Data":"f914ed9d0c2310782e3f5442bf1a28d331fac3c59cf65a4e749bd55fd4853437"} Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.864268 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.861620996 podStartE2EDuration="1m8.864236337s" podCreationTimestamp="2025-10-09 10:22:23 +0000 UTC" firstStartedPulling="2025-10-09 10:22:24.84032556 +0000 UTC m=+1030.908507316" lastFinishedPulling="2025-10-09 10:23:30.842940901 +0000 UTC m=+1096.911122657" observedRunningTime="2025-10-09 10:23:31.850191128 +0000 UTC m=+1097.918372884" watchObservedRunningTime="2025-10-09 10:23:31.864236337 +0000 UTC m=+1097.932418093" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.900583 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85494b87f-hlckn" podStartSLOduration=11.900554243 podStartE2EDuration="11.900554243s" podCreationTimestamp="2025-10-09 10:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:31.884386434 +0000 UTC m=+1097.952568190" watchObservedRunningTime="2025-10-09 10:23:31.900554243 +0000 UTC m=+1097.968735999" Oct 09 10:23:31 crc kubenswrapper[4923]: I1009 10:23:31.913255 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7bc74dbf64-gj94f" podStartSLOduration=10.013392784 podStartE2EDuration="13.913229293s" podCreationTimestamp="2025-10-09 10:23:18 +0000 UTC" firstStartedPulling="2025-10-09 10:23:19.59230068 +0000 UTC m=+1085.660482436" lastFinishedPulling="2025-10-09 10:23:23.492137189 +0000 UTC m=+1089.560318945" observedRunningTime="2025-10-09 10:23:31.900084259 +0000 UTC m=+1097.968266035" watchObservedRunningTime="2025-10-09 10:23:31.913229293 +0000 UTC m=+1097.981411039" Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.049486 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7c85c4b745-jf7ck"] Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.079274 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-855475c586-p22m4" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.092864 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7c85c4b745-jf7ck"] Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.187855 4923 scope.go:117] "RemoveContainer" containerID="230fb6ba9b99d46c0ba822b55f186e30ac29a85fef94e58740f5ad27d43191a3" Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.613549 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e39f0227-4155-4d15-9743-d81095c77c0d" path="/var/lib/kubelet/pods/e39f0227-4155-4d15-9743-d81095c77c0d/volumes" Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.875447 4923 generic.go:334] "Generic (PLEG): container finished" podID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerID="13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6" exitCode=0 Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.875913 4923 generic.go:334] "Generic (PLEG): container finished" podID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerID="4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad" exitCode=2 Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.875923 4923 generic.go:334] "Generic (PLEG): container finished" podID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerID="b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239" exitCode=0 Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.875497 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerDied","Data":"13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6"} Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.875995 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerDied","Data":"4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad"} Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.876024 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerDied","Data":"b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239"} Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.879267 4923 generic.go:334] "Generic (PLEG): container finished" podID="f9c24b4f-b830-494a-9789-745e176ffac6" containerID="b95dac3af77883b7de07aef6630527ed09e7cfba4c7092318dd041d2804ba167" exitCode=0 Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.879297 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-665d85ccb-zkkd2" event={"ID":"f9c24b4f-b830-494a-9789-745e176ffac6","Type":"ContainerDied","Data":"b95dac3af77883b7de07aef6630527ed09e7cfba4c7092318dd041d2804ba167"} Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.882723 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d29f311b-49a9-4251-b640-79dea462412f","Type":"ContainerStarted","Data":"bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118"} Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.882903 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d29f311b-49a9-4251-b640-79dea462412f" containerName="cinder-api-log" containerID="cri-o://1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c" gracePeriod=30 Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.883185 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.883456 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d29f311b-49a9-4251-b640-79dea462412f" containerName="cinder-api" containerID="cri-o://bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118" gracePeriod=30 Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.895047 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-9f849b6cd-m82rj" event={"ID":"45877ffa-4abb-489a-9bad-770cbb5f231a","Type":"ContainerStarted","Data":"32faf6032f770304cf0e731546a7b702e4de0a01c693d3a697a153f5f8d16dcd"} Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.895572 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.895608 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.912394 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=12.912369725 podStartE2EDuration="12.912369725s" podCreationTimestamp="2025-10-09 10:23:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:32.901930145 +0000 UTC m=+1098.970111901" watchObservedRunningTime="2025-10-09 10:23:32.912369725 +0000 UTC m=+1098.980551481" Oct 09 10:23:32 crc kubenswrapper[4923]: I1009 10:23:32.937064 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-9f849b6cd-m82rj" podStartSLOduration=7.937038347 podStartE2EDuration="7.937038347s" podCreationTimestamp="2025-10-09 10:23:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:32.933830759 +0000 UTC m=+1099.002012525" watchObservedRunningTime="2025-10-09 10:23:32.937038347 +0000 UTC m=+1099.005220103" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.279022 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.366332 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-config\") pod \"f9c24b4f-b830-494a-9789-745e176ffac6\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.366373 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-httpd-config\") pod \"f9c24b4f-b830-494a-9789-745e176ffac6\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.366451 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-combined-ca-bundle\") pod \"f9c24b4f-b830-494a-9789-745e176ffac6\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.366490 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9brq2\" (UniqueName: \"kubernetes.io/projected/f9c24b4f-b830-494a-9789-745e176ffac6-kube-api-access-9brq2\") pod \"f9c24b4f-b830-494a-9789-745e176ffac6\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.366602 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-ovndb-tls-certs\") pod \"f9c24b4f-b830-494a-9789-745e176ffac6\" (UID: \"f9c24b4f-b830-494a-9789-745e176ffac6\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.375691 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9c24b4f-b830-494a-9789-745e176ffac6-kube-api-access-9brq2" (OuterVolumeSpecName: "kube-api-access-9brq2") pod "f9c24b4f-b830-494a-9789-745e176ffac6" (UID: "f9c24b4f-b830-494a-9789-745e176ffac6"). InnerVolumeSpecName "kube-api-access-9brq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.379795 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "f9c24b4f-b830-494a-9789-745e176ffac6" (UID: "f9c24b4f-b830-494a-9789-745e176ffac6"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.438980 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f9c24b4f-b830-494a-9789-745e176ffac6" (UID: "f9c24b4f-b830-494a-9789-745e176ffac6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.452713 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-config" (OuterVolumeSpecName: "config") pod "f9c24b4f-b830-494a-9789-745e176ffac6" (UID: "f9c24b4f-b830-494a-9789-745e176ffac6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.467178 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "f9c24b4f-b830-494a-9789-745e176ffac6" (UID: "f9c24b4f-b830-494a-9789-745e176ffac6"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.471856 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.471888 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9brq2\" (UniqueName: \"kubernetes.io/projected/f9c24b4f-b830-494a-9789-745e176ffac6-kube-api-access-9brq2\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.471902 4923 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.471917 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.471929 4923 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/f9c24b4f-b830-494a-9789-745e176ffac6-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.654289 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.776617 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-scripts\") pod \"d29f311b-49a9-4251-b640-79dea462412f\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.776886 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d29f311b-49a9-4251-b640-79dea462412f-etc-machine-id\") pod \"d29f311b-49a9-4251-b640-79dea462412f\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.776922 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data-custom\") pod \"d29f311b-49a9-4251-b640-79dea462412f\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.776961 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29f311b-49a9-4251-b640-79dea462412f-logs\") pod \"d29f311b-49a9-4251-b640-79dea462412f\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.777026 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-combined-ca-bundle\") pod \"d29f311b-49a9-4251-b640-79dea462412f\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.777055 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data\") pod \"d29f311b-49a9-4251-b640-79dea462412f\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.777285 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkk8f\" (UniqueName: \"kubernetes.io/projected/d29f311b-49a9-4251-b640-79dea462412f-kube-api-access-kkk8f\") pod \"d29f311b-49a9-4251-b640-79dea462412f\" (UID: \"d29f311b-49a9-4251-b640-79dea462412f\") " Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.777578 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d29f311b-49a9-4251-b640-79dea462412f-logs" (OuterVolumeSpecName: "logs") pod "d29f311b-49a9-4251-b640-79dea462412f" (UID: "d29f311b-49a9-4251-b640-79dea462412f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.777046 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d29f311b-49a9-4251-b640-79dea462412f-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d29f311b-49a9-4251-b640-79dea462412f" (UID: "d29f311b-49a9-4251-b640-79dea462412f"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.779835 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d29f311b-49a9-4251-b640-79dea462412f-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.779865 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d29f311b-49a9-4251-b640-79dea462412f-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.782395 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d29f311b-49a9-4251-b640-79dea462412f-kube-api-access-kkk8f" (OuterVolumeSpecName: "kube-api-access-kkk8f") pod "d29f311b-49a9-4251-b640-79dea462412f" (UID: "d29f311b-49a9-4251-b640-79dea462412f"). InnerVolumeSpecName "kube-api-access-kkk8f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.799939 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d29f311b-49a9-4251-b640-79dea462412f" (UID: "d29f311b-49a9-4251-b640-79dea462412f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.799999 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-scripts" (OuterVolumeSpecName: "scripts") pod "d29f311b-49a9-4251-b640-79dea462412f" (UID: "d29f311b-49a9-4251-b640-79dea462412f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.804833 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d29f311b-49a9-4251-b640-79dea462412f" (UID: "d29f311b-49a9-4251-b640-79dea462412f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.824435 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data" (OuterVolumeSpecName: "config-data") pod "d29f311b-49a9-4251-b640-79dea462412f" (UID: "d29f311b-49a9-4251-b640-79dea462412f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.881868 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.881899 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.881909 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.881918 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d29f311b-49a9-4251-b640-79dea462412f-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.881927 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkk8f\" (UniqueName: \"kubernetes.io/projected/d29f311b-49a9-4251-b640-79dea462412f-kube-api-access-kkk8f\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.906505 4923 generic.go:334] "Generic (PLEG): container finished" podID="d29f311b-49a9-4251-b640-79dea462412f" containerID="bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118" exitCode=0 Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.906728 4923 generic.go:334] "Generic (PLEG): container finished" podID="d29f311b-49a9-4251-b640-79dea462412f" containerID="1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c" exitCode=143 Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.906605 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d29f311b-49a9-4251-b640-79dea462412f","Type":"ContainerDied","Data":"bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118"} Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.906576 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.907102 4923 scope.go:117] "RemoveContainer" containerID="bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118" Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.907049 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d29f311b-49a9-4251-b640-79dea462412f","Type":"ContainerDied","Data":"1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c"} Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.907342 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d29f311b-49a9-4251-b640-79dea462412f","Type":"ContainerDied","Data":"0a4299f657e05a8e0357268cc573bbfc564a45b92474ef463f4be735b4b81403"} Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.909050 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-665d85ccb-zkkd2" event={"ID":"f9c24b4f-b830-494a-9789-745e176ffac6","Type":"ContainerDied","Data":"5474096c79a28be441c85f6a1180a8a74a7885f18447189d15cbb555cb9b5292"} Oct 09 10:23:33 crc kubenswrapper[4923]: I1009 10:23:33.909087 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-665d85ccb-zkkd2" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.072082 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.090521 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.092149 4923 scope.go:117] "RemoveContainer" containerID="1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.115528 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-665d85ccb-zkkd2"] Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.127684 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-665d85ccb-zkkd2"] Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.151522 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 09 10:23:34 crc kubenswrapper[4923]: E1009 10:23:34.152116 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c24b4f-b830-494a-9789-745e176ffac6" containerName="neutron-httpd" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.152139 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c24b4f-b830-494a-9789-745e176ffac6" containerName="neutron-httpd" Oct 09 10:23:34 crc kubenswrapper[4923]: E1009 10:23:34.152151 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29f311b-49a9-4251-b640-79dea462412f" containerName="cinder-api" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.152158 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29f311b-49a9-4251-b640-79dea462412f" containerName="cinder-api" Oct 09 10:23:34 crc kubenswrapper[4923]: E1009 10:23:34.152171 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e39f0227-4155-4d15-9743-d81095c77c0d" containerName="horizon-log" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.152181 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e39f0227-4155-4d15-9743-d81095c77c0d" containerName="horizon-log" Oct 09 10:23:34 crc kubenswrapper[4923]: E1009 10:23:34.152198 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e39f0227-4155-4d15-9743-d81095c77c0d" containerName="horizon" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.152206 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e39f0227-4155-4d15-9743-d81095c77c0d" containerName="horizon" Oct 09 10:23:34 crc kubenswrapper[4923]: E1009 10:23:34.152227 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29f311b-49a9-4251-b640-79dea462412f" containerName="cinder-api-log" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.152237 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29f311b-49a9-4251-b640-79dea462412f" containerName="cinder-api-log" Oct 09 10:23:34 crc kubenswrapper[4923]: E1009 10:23:34.152266 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9c24b4f-b830-494a-9789-745e176ffac6" containerName="neutron-api" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.152273 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9c24b4f-b830-494a-9789-745e176ffac6" containerName="neutron-api" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.152610 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e39f0227-4155-4d15-9743-d81095c77c0d" containerName="horizon" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.153380 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e39f0227-4155-4d15-9743-d81095c77c0d" containerName="horizon-log" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.153399 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c24b4f-b830-494a-9789-745e176ffac6" containerName="neutron-httpd" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.153410 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d29f311b-49a9-4251-b640-79dea462412f" containerName="cinder-api-log" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.153424 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9c24b4f-b830-494a-9789-745e176ffac6" containerName="neutron-api" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.153442 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d29f311b-49a9-4251-b640-79dea462412f" containerName="cinder-api" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.155656 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.160579 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.161670 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.165845 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.166473 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.166476 4923 scope.go:117] "RemoveContainer" containerID="bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118" Oct 09 10:23:34 crc kubenswrapper[4923]: E1009 10:23:34.172801 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118\": container with ID starting with bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118 not found: ID does not exist" containerID="bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.172845 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118"} err="failed to get container status \"bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118\": rpc error: code = NotFound desc = could not find container \"bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118\": container with ID starting with bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118 not found: ID does not exist" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.172888 4923 scope.go:117] "RemoveContainer" containerID="1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c" Oct 09 10:23:34 crc kubenswrapper[4923]: E1009 10:23:34.173556 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c\": container with ID starting with 1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c not found: ID does not exist" containerID="1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.173587 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c"} err="failed to get container status \"1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c\": rpc error: code = NotFound desc = could not find container \"1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c\": container with ID starting with 1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c not found: ID does not exist" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.173605 4923 scope.go:117] "RemoveContainer" containerID="bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.173887 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118"} err="failed to get container status \"bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118\": rpc error: code = NotFound desc = could not find container \"bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118\": container with ID starting with bca9c6cc159b949393e0cbefd405389d169ef0c88014e53421c642b4a62e5118 not found: ID does not exist" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.173912 4923 scope.go:117] "RemoveContainer" containerID="1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.177031 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c"} err="failed to get container status \"1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c\": rpc error: code = NotFound desc = could not find container \"1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c\": container with ID starting with 1d1560ebc82aee0ac2a63e90ce7b07c3e89d7cbc15adf00ad966bf7ecb3d078c not found: ID does not exist" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.177058 4923 scope.go:117] "RemoveContainer" containerID="86bf1e0b381981b4936ffb857b09e0bfb1091b6d8d9c1a65c1150d5f4cfaa492" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.208065 4923 scope.go:117] "RemoveContainer" containerID="b95dac3af77883b7de07aef6630527ed09e7cfba4c7092318dd041d2804ba167" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.277287 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.291351 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-logs\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.291731 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-public-tls-certs\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.292031 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-scripts\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.292081 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-config-data-custom\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.292192 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.292245 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.292272 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-etc-machine-id\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.292980 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g8lm\" (UniqueName: \"kubernetes.io/projected/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-kube-api-access-5g8lm\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.293207 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-config-data\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.395034 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc87w\" (UniqueName: \"kubernetes.io/projected/66d50d55-3a5f-4027-aa14-3abf8d0e3709-kube-api-access-lc87w\") pod \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.395170 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-scripts\") pod \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.395354 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-log-httpd\") pod \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.395381 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-combined-ca-bundle\") pod \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.395498 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-sg-core-conf-yaml\") pod \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.395601 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-config-data\") pod \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.395629 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-run-httpd\") pod \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\" (UID: \"66d50d55-3a5f-4027-aa14-3abf8d0e3709\") " Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.395977 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-public-tls-certs\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396046 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-scripts\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396080 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-config-data-custom\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396109 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396133 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396151 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-etc-machine-id\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396250 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g8lm\" (UniqueName: \"kubernetes.io/projected/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-kube-api-access-5g8lm\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396304 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-config-data\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396325 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-logs\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396320 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "66d50d55-3a5f-4027-aa14-3abf8d0e3709" (UID: "66d50d55-3a5f-4027-aa14-3abf8d0e3709"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396541 4923 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396843 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-logs\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396898 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-etc-machine-id\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.396938 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "66d50d55-3a5f-4027-aa14-3abf8d0e3709" (UID: "66d50d55-3a5f-4027-aa14-3abf8d0e3709"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.402090 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-public-tls-certs\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.404106 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-scripts" (OuterVolumeSpecName: "scripts") pod "66d50d55-3a5f-4027-aa14-3abf8d0e3709" (UID: "66d50d55-3a5f-4027-aa14-3abf8d0e3709"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.405016 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.410983 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d50d55-3a5f-4027-aa14-3abf8d0e3709-kube-api-access-lc87w" (OuterVolumeSpecName: "kube-api-access-lc87w") pod "66d50d55-3a5f-4027-aa14-3abf8d0e3709" (UID: "66d50d55-3a5f-4027-aa14-3abf8d0e3709"). InnerVolumeSpecName "kube-api-access-lc87w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.412389 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-config-data-custom\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.413232 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-config-data\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.413640 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-scripts\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.415630 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g8lm\" (UniqueName: \"kubernetes.io/projected/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-kube-api-access-5g8lm\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.415958 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62\") " pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.437104 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "66d50d55-3a5f-4027-aa14-3abf8d0e3709" (UID: "66d50d55-3a5f-4027-aa14-3abf8d0e3709"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.486673 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66d50d55-3a5f-4027-aa14-3abf8d0e3709" (UID: "66d50d55-3a5f-4027-aa14-3abf8d0e3709"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.498526 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.498672 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.498697 4923 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.498716 4923 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/66d50d55-3a5f-4027-aa14-3abf8d0e3709-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.498726 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc87w\" (UniqueName: \"kubernetes.io/projected/66d50d55-3a5f-4027-aa14-3abf8d0e3709-kube-api-access-lc87w\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.498736 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.502306 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-config-data" (OuterVolumeSpecName: "config-data") pod "66d50d55-3a5f-4027-aa14-3abf8d0e3709" (UID: "66d50d55-3a5f-4027-aa14-3abf8d0e3709"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.601359 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66d50d55-3a5f-4027-aa14-3abf8d0e3709-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.628679 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d29f311b-49a9-4251-b640-79dea462412f" path="/var/lib/kubelet/pods/d29f311b-49a9-4251-b640-79dea462412f/volumes" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.629601 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9c24b4f-b830-494a-9789-745e176ffac6" path="/var/lib/kubelet/pods/f9c24b4f-b830-494a-9789-745e176ffac6/volumes" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.927431 4923 generic.go:334] "Generic (PLEG): container finished" podID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerID="5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f" exitCode=0 Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.927461 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerDied","Data":"5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f"} Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.927917 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"66d50d55-3a5f-4027-aa14-3abf8d0e3709","Type":"ContainerDied","Data":"77b40526dd4b1429861b03227e663d440f5dc657b8378055e09ca5e89731abc4"} Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.927486 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.927947 4923 scope.go:117] "RemoveContainer" containerID="13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.962209 4923 scope.go:117] "RemoveContainer" containerID="4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad" Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.969229 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:23:34 crc kubenswrapper[4923]: I1009 10:23:34.993257 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.001115 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 09 10:23:35 crc kubenswrapper[4923]: W1009 10:23:35.001955 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod07de3fcd_6ca3_4c5a_a57b_97a51ffe6d62.slice/crio-69831153bde6514a3d412bf7a4df2bc86d933f8551b88590401d63af2a8793d7 WatchSource:0}: Error finding container 69831153bde6514a3d412bf7a4df2bc86d933f8551b88590401d63af2a8793d7: Status 404 returned error can't find the container with id 69831153bde6514a3d412bf7a4df2bc86d933f8551b88590401d63af2a8793d7 Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.016387 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:23:35 crc kubenswrapper[4923]: E1009 10:23:35.016855 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="ceilometer-notification-agent" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.016876 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="ceilometer-notification-agent" Oct 09 10:23:35 crc kubenswrapper[4923]: E1009 10:23:35.016905 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="proxy-httpd" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.016912 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="proxy-httpd" Oct 09 10:23:35 crc kubenswrapper[4923]: E1009 10:23:35.016922 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="sg-core" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.016928 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="sg-core" Oct 09 10:23:35 crc kubenswrapper[4923]: E1009 10:23:35.016951 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="ceilometer-central-agent" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.016957 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="ceilometer-central-agent" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.017182 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="ceilometer-central-agent" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.017203 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="proxy-httpd" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.017214 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="sg-core" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.017236 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" containerName="ceilometer-notification-agent" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.019265 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.026913 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.027266 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.030029 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.031954 4923 scope.go:117] "RemoveContainer" containerID="5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.071410 4923 scope.go:117] "RemoveContainer" containerID="b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.096667 4923 scope.go:117] "RemoveContainer" containerID="13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6" Oct 09 10:23:35 crc kubenswrapper[4923]: E1009 10:23:35.097233 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6\": container with ID starting with 13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6 not found: ID does not exist" containerID="13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.097275 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6"} err="failed to get container status \"13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6\": rpc error: code = NotFound desc = could not find container \"13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6\": container with ID starting with 13eaf9b4d560d3bb567f818053df8547b24cfa9687e378a8084c79c711ca33b6 not found: ID does not exist" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.097307 4923 scope.go:117] "RemoveContainer" containerID="4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad" Oct 09 10:23:35 crc kubenswrapper[4923]: E1009 10:23:35.097663 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad\": container with ID starting with 4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad not found: ID does not exist" containerID="4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.097690 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad"} err="failed to get container status \"4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad\": rpc error: code = NotFound desc = could not find container \"4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad\": container with ID starting with 4a25955f328e7f84c70a71c6332cd06a7c5f774e9bd3e462105e29d991811dad not found: ID does not exist" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.097710 4923 scope.go:117] "RemoveContainer" containerID="5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f" Oct 09 10:23:35 crc kubenswrapper[4923]: E1009 10:23:35.098371 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f\": container with ID starting with 5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f not found: ID does not exist" containerID="5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.098399 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f"} err="failed to get container status \"5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f\": rpc error: code = NotFound desc = could not find container \"5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f\": container with ID starting with 5c0ae19dd56088af8b29dbf77e04c6b2dd33605312b1d179d31d3b7ada57b72f not found: ID does not exist" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.098414 4923 scope.go:117] "RemoveContainer" containerID="b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239" Oct 09 10:23:35 crc kubenswrapper[4923]: E1009 10:23:35.098850 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239\": container with ID starting with b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239 not found: ID does not exist" containerID="b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.098874 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239"} err="failed to get container status \"b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239\": rpc error: code = NotFound desc = could not find container \"b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239\": container with ID starting with b0b76f10a392617f7a506f5692584c96a640d59893a648fb58e12a1601774239 not found: ID does not exist" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.114446 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-scripts\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.114516 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-run-httpd\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.114568 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-config-data\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.114639 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.114669 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxstj\" (UniqueName: \"kubernetes.io/projected/5598fa65-0ab7-4995-b47c-d144192572bc-kube-api-access-qxstj\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.114800 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-log-httpd\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.114865 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.216623 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.216691 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxstj\" (UniqueName: \"kubernetes.io/projected/5598fa65-0ab7-4995-b47c-d144192572bc-kube-api-access-qxstj\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.216803 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-log-httpd\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.216847 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.216882 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-run-httpd\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.216907 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-scripts\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.216949 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-config-data\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.218401 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-run-httpd\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.218535 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-log-httpd\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.224845 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.225483 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-config-data\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.226638 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.227122 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-scripts\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.238159 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxstj\" (UniqueName: \"kubernetes.io/projected/5598fa65-0ab7-4995-b47c-d144192572bc-kube-api-access-qxstj\") pod \"ceilometer-0\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.345313 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.811227 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:23:35 crc kubenswrapper[4923]: W1009 10:23:35.822776 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5598fa65_0ab7_4995_b47c_d144192572bc.slice/crio-5daba828d09fa5281e4eac51650ad1cbd95d5910b1d39ab9e900c8d9571eecb9 WatchSource:0}: Error finding container 5daba828d09fa5281e4eac51650ad1cbd95d5910b1d39ab9e900c8d9571eecb9: Status 404 returned error can't find the container with id 5daba828d09fa5281e4eac51650ad1cbd95d5910b1d39ab9e900c8d9571eecb9 Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.945987 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerStarted","Data":"5daba828d09fa5281e4eac51650ad1cbd95d5910b1d39ab9e900c8d9571eecb9"} Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.947515 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62","Type":"ContainerStarted","Data":"b3a03ea9ce63b35d0716d09a4d90470902b82bf4dfedc8737b0fc8996290e62d"} Oct 09 10:23:35 crc kubenswrapper[4923]: I1009 10:23:35.947571 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62","Type":"ContainerStarted","Data":"69831153bde6514a3d412bf7a4df2bc86d933f8551b88590401d63af2a8793d7"} Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.007601 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.127000 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.241156 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh"] Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.248929 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" podUID="8395d8be-e7bb-4136-a47a-e61770e52b99" containerName="dnsmasq-dns" containerID="cri-o://36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1" gracePeriod=10 Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.266677 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.639052 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66d50d55-3a5f-4027-aa14-3abf8d0e3709" path="/var/lib/kubelet/pods/66d50d55-3a5f-4027-aa14-3abf8d0e3709/volumes" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.765138 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.864672 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp6cg\" (UniqueName: \"kubernetes.io/projected/8395d8be-e7bb-4136-a47a-e61770e52b99-kube-api-access-zp6cg\") pod \"8395d8be-e7bb-4136-a47a-e61770e52b99\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.864732 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-config\") pod \"8395d8be-e7bb-4136-a47a-e61770e52b99\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.864946 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-nb\") pod \"8395d8be-e7bb-4136-a47a-e61770e52b99\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.865060 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-sb\") pod \"8395d8be-e7bb-4136-a47a-e61770e52b99\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.865115 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-dns-svc\") pod \"8395d8be-e7bb-4136-a47a-e61770e52b99\" (UID: \"8395d8be-e7bb-4136-a47a-e61770e52b99\") " Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.882138 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8395d8be-e7bb-4136-a47a-e61770e52b99-kube-api-access-zp6cg" (OuterVolumeSpecName: "kube-api-access-zp6cg") pod "8395d8be-e7bb-4136-a47a-e61770e52b99" (UID: "8395d8be-e7bb-4136-a47a-e61770e52b99"). InnerVolumeSpecName "kube-api-access-zp6cg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.920470 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8395d8be-e7bb-4136-a47a-e61770e52b99" (UID: "8395d8be-e7bb-4136-a47a-e61770e52b99"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.922072 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8395d8be-e7bb-4136-a47a-e61770e52b99" (UID: "8395d8be-e7bb-4136-a47a-e61770e52b99"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.928406 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-config" (OuterVolumeSpecName: "config") pod "8395d8be-e7bb-4136-a47a-e61770e52b99" (UID: "8395d8be-e7bb-4136-a47a-e61770e52b99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.943232 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8395d8be-e7bb-4136-a47a-e61770e52b99" (UID: "8395d8be-e7bb-4136-a47a-e61770e52b99"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.964617 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerStarted","Data":"c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369"} Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.968788 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.968872 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.968886 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.968895 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp6cg\" (UniqueName: \"kubernetes.io/projected/8395d8be-e7bb-4136-a47a-e61770e52b99-kube-api-access-zp6cg\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.968907 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8395d8be-e7bb-4136-a47a-e61770e52b99-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.975769 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62","Type":"ContainerStarted","Data":"e4701472edd835fb325e3de1b35d5fd0139b7f3d02255e8055f280bb0d344835"} Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.976241 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.981128 4923 generic.go:334] "Generic (PLEG): container finished" podID="8395d8be-e7bb-4136-a47a-e61770e52b99" containerID="36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1" exitCode=0 Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.981243 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" event={"ID":"8395d8be-e7bb-4136-a47a-e61770e52b99","Type":"ContainerDied","Data":"36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1"} Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.981328 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.981407 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh" event={"ID":"8395d8be-e7bb-4136-a47a-e61770e52b99","Type":"ContainerDied","Data":"053e3e68ab8f57d185a55cb5951c4aec6f941420cffcbff32152b62dd8b2b637"} Oct 09 10:23:36 crc kubenswrapper[4923]: I1009 10:23:36.981438 4923 scope.go:117] "RemoveContainer" containerID="36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1" Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.019091 4923 scope.go:117] "RemoveContainer" containerID="4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b" Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.043207 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.043117126 podStartE2EDuration="3.043117126s" podCreationTimestamp="2025-10-09 10:23:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:37.001709539 +0000 UTC m=+1103.069891315" watchObservedRunningTime="2025-10-09 10:23:37.043117126 +0000 UTC m=+1103.111298882" Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.059404 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.065681 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh"] Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.073653 4923 scope.go:117] "RemoveContainer" containerID="36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1" Oct 09 10:23:37 crc kubenswrapper[4923]: E1009 10:23:37.076987 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1\": container with ID starting with 36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1 not found: ID does not exist" containerID="36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1" Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.077037 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1"} err="failed to get container status \"36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1\": rpc error: code = NotFound desc = could not find container \"36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1\": container with ID starting with 36191af1ead2618f0b1c8881122f595e5b0d314f39dcc32a2ea0c9de50c2e2a1 not found: ID does not exist" Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.077070 4923 scope.go:117] "RemoveContainer" containerID="4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b" Oct 09 10:23:37 crc kubenswrapper[4923]: E1009 10:23:37.078207 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b\": container with ID starting with 4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b not found: ID does not exist" containerID="4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b" Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.078263 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b"} err="failed to get container status \"4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b\": rpc error: code = NotFound desc = could not find container \"4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b\": container with ID starting with 4152e97e70927032a7dc1f5b1c7e774e4ea97f0ecbcc3e0fb6dc910031fef84b not found: ID does not exist" Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.079489 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f7d8dc7ff-4jtmh"] Oct 09 10:23:37 crc kubenswrapper[4923]: E1009 10:23:37.323295 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf250b2a7_fb69_4c40_b186_238f74472a44.slice/crio-3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a\": RecentStats: unable to find data in memory cache]" Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.992664 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerStarted","Data":"39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec"} Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.994055 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="72e46628-da39-4445-a4e2-0d372fce2716" containerName="cinder-scheduler" containerID="cri-o://f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47" gracePeriod=30 Oct 09 10:23:37 crc kubenswrapper[4923]: I1009 10:23:37.994170 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="72e46628-da39-4445-a4e2-0d372fce2716" containerName="probe" containerID="cri-o://1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9" gracePeriod=30 Oct 09 10:23:38 crc kubenswrapper[4923]: I1009 10:23:38.619863 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8395d8be-e7bb-4136-a47a-e61770e52b99" path="/var/lib/kubelet/pods/8395d8be-e7bb-4136-a47a-e61770e52b99/volumes" Oct 09 10:23:39 crc kubenswrapper[4923]: I1009 10:23:39.006795 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerStarted","Data":"ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0"} Oct 09 10:23:40 crc kubenswrapper[4923]: I1009 10:23:40.022341 4923 generic.go:334] "Generic (PLEG): container finished" podID="72e46628-da39-4445-a4e2-0d372fce2716" containerID="1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9" exitCode=0 Oct 09 10:23:40 crc kubenswrapper[4923]: I1009 10:23:40.023952 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72e46628-da39-4445-a4e2-0d372fce2716","Type":"ContainerDied","Data":"1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9"} Oct 09 10:23:41 crc kubenswrapper[4923]: I1009 10:23:41.037200 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerStarted","Data":"0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb"} Oct 09 10:23:41 crc kubenswrapper[4923]: I1009 10:23:41.038973 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 10:23:41 crc kubenswrapper[4923]: I1009 10:23:41.084896 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.741747267 podStartE2EDuration="7.084861566s" podCreationTimestamp="2025-10-09 10:23:34 +0000 UTC" firstStartedPulling="2025-10-09 10:23:35.825242487 +0000 UTC m=+1101.893424253" lastFinishedPulling="2025-10-09 10:23:40.168356796 +0000 UTC m=+1106.236538552" observedRunningTime="2025-10-09 10:23:41.071223841 +0000 UTC m=+1107.139405617" watchObservedRunningTime="2025-10-09 10:23:41.084861566 +0000 UTC m=+1107.153043312" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.038291 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.057165 4923 generic.go:334] "Generic (PLEG): container finished" podID="72e46628-da39-4445-a4e2-0d372fce2716" containerID="f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47" exitCode=0 Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.057242 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72e46628-da39-4445-a4e2-0d372fce2716","Type":"ContainerDied","Data":"f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47"} Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.057310 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"72e46628-da39-4445-a4e2-0d372fce2716","Type":"ContainerDied","Data":"63586d2cbbcfc7bb7be2346d3733da333bb6f9715c18e6e41a444358bbc103b2"} Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.057306 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.057331 4923 scope.go:117] "RemoveContainer" containerID="1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.082864 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-855475c586-p22m4" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.115044 4923 scope.go:117] "RemoveContainer" containerID="f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.172056 4923 scope.go:117] "RemoveContainer" containerID="1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9" Oct 09 10:23:42 crc kubenswrapper[4923]: E1009 10:23:42.173413 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9\": container with ID starting with 1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9 not found: ID does not exist" containerID="1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.173461 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9"} err="failed to get container status \"1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9\": rpc error: code = NotFound desc = could not find container \"1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9\": container with ID starting with 1e962a972c6d8863b40dd1e25c1a97394fd66a0fac4b67fed8d41c0681b0fdd9 not found: ID does not exist" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.173498 4923 scope.go:117] "RemoveContainer" containerID="f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47" Oct 09 10:23:42 crc kubenswrapper[4923]: E1009 10:23:42.174264 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47\": container with ID starting with f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47 not found: ID does not exist" containerID="f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.174413 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47"} err="failed to get container status \"f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47\": rpc error: code = NotFound desc = could not find container \"f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47\": container with ID starting with f6bc7c5890dac308af5697dcc48f65507f97ffbe12fdcd5ea73ab274bc7cdb47 not found: ID does not exist" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.180010 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-scripts\") pod \"72e46628-da39-4445-a4e2-0d372fce2716\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.180303 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-combined-ca-bundle\") pod \"72e46628-da39-4445-a4e2-0d372fce2716\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.180684 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data-custom\") pod \"72e46628-da39-4445-a4e2-0d372fce2716\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.180776 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgvw8\" (UniqueName: \"kubernetes.io/projected/72e46628-da39-4445-a4e2-0d372fce2716-kube-api-access-hgvw8\") pod \"72e46628-da39-4445-a4e2-0d372fce2716\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.180824 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72e46628-da39-4445-a4e2-0d372fce2716-etc-machine-id\") pod \"72e46628-da39-4445-a4e2-0d372fce2716\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.180945 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data\") pod \"72e46628-da39-4445-a4e2-0d372fce2716\" (UID: \"72e46628-da39-4445-a4e2-0d372fce2716\") " Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.181153 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/72e46628-da39-4445-a4e2-0d372fce2716-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "72e46628-da39-4445-a4e2-0d372fce2716" (UID: "72e46628-da39-4445-a4e2-0d372fce2716"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.181893 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/72e46628-da39-4445-a4e2-0d372fce2716-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.191000 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72e46628-da39-4445-a4e2-0d372fce2716-kube-api-access-hgvw8" (OuterVolumeSpecName: "kube-api-access-hgvw8") pod "72e46628-da39-4445-a4e2-0d372fce2716" (UID: "72e46628-da39-4445-a4e2-0d372fce2716"). InnerVolumeSpecName "kube-api-access-hgvw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.191382 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "72e46628-da39-4445-a4e2-0d372fce2716" (UID: "72e46628-da39-4445-a4e2-0d372fce2716"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.209995 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.215955 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-scripts" (OuterVolumeSpecName: "scripts") pod "72e46628-da39-4445-a4e2-0d372fce2716" (UID: "72e46628-da39-4445-a4e2-0d372fce2716"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.254584 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "72e46628-da39-4445-a4e2-0d372fce2716" (UID: "72e46628-da39-4445-a4e2-0d372fce2716"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.283290 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-9f849b6cd-m82rj" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.283716 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.283798 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgvw8\" (UniqueName: \"kubernetes.io/projected/72e46628-da39-4445-a4e2-0d372fce2716-kube-api-access-hgvw8\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.283814 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.283824 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.291932 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data" (OuterVolumeSpecName: "config-data") pod "72e46628-da39-4445-a4e2-0d372fce2716" (UID: "72e46628-da39-4445-a4e2-0d372fce2716"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.369763 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6744bb4d8-9hkdw"] Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.370053 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6744bb4d8-9hkdw" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api-log" containerID="cri-o://7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad" gracePeriod=30 Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.370508 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6744bb4d8-9hkdw" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api" containerID="cri-o://0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8" gracePeriod=30 Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.390185 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/72e46628-da39-4445-a4e2-0d372fce2716-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.429822 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.444651 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.467619 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 10:23:42 crc kubenswrapper[4923]: E1009 10:23:42.468128 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8395d8be-e7bb-4136-a47a-e61770e52b99" containerName="init" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.468153 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8395d8be-e7bb-4136-a47a-e61770e52b99" containerName="init" Oct 09 10:23:42 crc kubenswrapper[4923]: E1009 10:23:42.468164 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e46628-da39-4445-a4e2-0d372fce2716" containerName="cinder-scheduler" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.468171 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e46628-da39-4445-a4e2-0d372fce2716" containerName="cinder-scheduler" Oct 09 10:23:42 crc kubenswrapper[4923]: E1009 10:23:42.468193 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8395d8be-e7bb-4136-a47a-e61770e52b99" containerName="dnsmasq-dns" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.468202 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8395d8be-e7bb-4136-a47a-e61770e52b99" containerName="dnsmasq-dns" Oct 09 10:23:42 crc kubenswrapper[4923]: E1009 10:23:42.468228 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72e46628-da39-4445-a4e2-0d372fce2716" containerName="probe" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.468234 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="72e46628-da39-4445-a4e2-0d372fce2716" containerName="probe" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.474586 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="8395d8be-e7bb-4136-a47a-e61770e52b99" containerName="dnsmasq-dns" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.474626 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e46628-da39-4445-a4e2-0d372fce2716" containerName="probe" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.474654 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="72e46628-da39-4445-a4e2-0d372fce2716" containerName="cinder-scheduler" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.476174 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.480933 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.517595 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.595315 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.595414 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e99cc2d5-9514-4cd1-9fb7-708774861057-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.595463 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.595801 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-config-data\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.596134 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-st9j6\" (UniqueName: \"kubernetes.io/projected/e99cc2d5-9514-4cd1-9fb7-708774861057-kube-api-access-st9j6\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.596243 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-scripts\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.612931 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72e46628-da39-4445-a4e2-0d372fce2716" path="/var/lib/kubelet/pods/72e46628-da39-4445-a4e2-0d372fce2716/volumes" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.706081 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-config-data\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.706211 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-st9j6\" (UniqueName: \"kubernetes.io/projected/e99cc2d5-9514-4cd1-9fb7-708774861057-kube-api-access-st9j6\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.706241 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-scripts\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.706300 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.706349 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e99cc2d5-9514-4cd1-9fb7-708774861057-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.706379 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.706939 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e99cc2d5-9514-4cd1-9fb7-708774861057-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.724154 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-config-data\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.739181 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.753120 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-st9j6\" (UniqueName: \"kubernetes.io/projected/e99cc2d5-9514-4cd1-9fb7-708774861057-kube-api-access-st9j6\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.753206 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.762018 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e99cc2d5-9514-4cd1-9fb7-708774861057-scripts\") pod \"cinder-scheduler-0\" (UID: \"e99cc2d5-9514-4cd1-9fb7-708774861057\") " pod="openstack/cinder-scheduler-0" Oct 09 10:23:42 crc kubenswrapper[4923]: I1009 10:23:42.843447 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 10:23:43 crc kubenswrapper[4923]: I1009 10:23:43.082230 4923 generic.go:334] "Generic (PLEG): container finished" podID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerID="7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad" exitCode=143 Oct 09 10:23:43 crc kubenswrapper[4923]: I1009 10:23:43.083002 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6744bb4d8-9hkdw" event={"ID":"e4d125b1-feb1-43f8-94ad-417f4e163822","Type":"ContainerDied","Data":"7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad"} Oct 09 10:23:43 crc kubenswrapper[4923]: I1009 10:23:43.364443 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 10:23:43 crc kubenswrapper[4923]: W1009 10:23:43.367133 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode99cc2d5_9514_4cd1_9fb7_708774861057.slice/crio-3ae010bddcd203b422875d028d55edeaf84b1364604ec788f321b55ef7a20e43 WatchSource:0}: Error finding container 3ae010bddcd203b422875d028d55edeaf84b1364604ec788f321b55ef7a20e43: Status 404 returned error can't find the container with id 3ae010bddcd203b422875d028d55edeaf84b1364604ec788f321b55ef7a20e43 Oct 09 10:23:44 crc kubenswrapper[4923]: I1009 10:23:44.119893 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e99cc2d5-9514-4cd1-9fb7-708774861057","Type":"ContainerStarted","Data":"dcddbf25621081cd553d837fcceb086c687d7d9f92982d739bfa70c1c8a3ac9f"} Oct 09 10:23:44 crc kubenswrapper[4923]: I1009 10:23:44.120522 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e99cc2d5-9514-4cd1-9fb7-708774861057","Type":"ContainerStarted","Data":"3ae010bddcd203b422875d028d55edeaf84b1364604ec788f321b55ef7a20e43"} Oct 09 10:23:45 crc kubenswrapper[4923]: I1009 10:23:45.132941 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e99cc2d5-9514-4cd1-9fb7-708774861057","Type":"ContainerStarted","Data":"fcc9c043bf1ba4ec041acba0919a459874f2dcf99d53873864e85259d8110d81"} Oct 09 10:23:45 crc kubenswrapper[4923]: I1009 10:23:45.176034 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.176011113 podStartE2EDuration="3.176011113s" podCreationTimestamp="2025-10-09 10:23:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:23:45.15586747 +0000 UTC m=+1111.224049236" watchObservedRunningTime="2025-10-09 10:23:45.176011113 +0000 UTC m=+1111.244192879" Oct 09 10:23:45 crc kubenswrapper[4923]: I1009 10:23:45.626137 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6744bb4d8-9hkdw" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:44890->10.217.0.152:9311: read: connection reset by peer" Oct 09 10:23:45 crc kubenswrapper[4923]: I1009 10:23:45.626858 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6744bb4d8-9hkdw" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.152:9311/healthcheck\": read tcp 10.217.0.2:44878->10.217.0.152:9311: read: connection reset by peer" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.082347 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.147025 4923 generic.go:334] "Generic (PLEG): container finished" podID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerID="0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8" exitCode=0 Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.147110 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6744bb4d8-9hkdw" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.147104 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6744bb4d8-9hkdw" event={"ID":"e4d125b1-feb1-43f8-94ad-417f4e163822","Type":"ContainerDied","Data":"0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8"} Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.147274 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6744bb4d8-9hkdw" event={"ID":"e4d125b1-feb1-43f8-94ad-417f4e163822","Type":"ContainerDied","Data":"715b146c872f0b5a9d2f682807657969f373dfb5f2015d1b418b54e3b7d9bff3"} Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.147305 4923 scope.go:117] "RemoveContainer" containerID="0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.182800 4923 scope.go:117] "RemoveContainer" containerID="7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.188501 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4d125b1-feb1-43f8-94ad-417f4e163822-logs\") pod \"e4d125b1-feb1-43f8-94ad-417f4e163822\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.188596 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjzzt\" (UniqueName: \"kubernetes.io/projected/e4d125b1-feb1-43f8-94ad-417f4e163822-kube-api-access-zjzzt\") pod \"e4d125b1-feb1-43f8-94ad-417f4e163822\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.188721 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data\") pod \"e4d125b1-feb1-43f8-94ad-417f4e163822\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.188782 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data-custom\") pod \"e4d125b1-feb1-43f8-94ad-417f4e163822\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.188826 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-combined-ca-bundle\") pod \"e4d125b1-feb1-43f8-94ad-417f4e163822\" (UID: \"e4d125b1-feb1-43f8-94ad-417f4e163822\") " Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.190864 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4d125b1-feb1-43f8-94ad-417f4e163822-logs" (OuterVolumeSpecName: "logs") pod "e4d125b1-feb1-43f8-94ad-417f4e163822" (UID: "e4d125b1-feb1-43f8-94ad-417f4e163822"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.197449 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e4d125b1-feb1-43f8-94ad-417f4e163822" (UID: "e4d125b1-feb1-43f8-94ad-417f4e163822"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.198194 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4d125b1-feb1-43f8-94ad-417f4e163822-kube-api-access-zjzzt" (OuterVolumeSpecName: "kube-api-access-zjzzt") pod "e4d125b1-feb1-43f8-94ad-417f4e163822" (UID: "e4d125b1-feb1-43f8-94ad-417f4e163822"). InnerVolumeSpecName "kube-api-access-zjzzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.207513 4923 scope.go:117] "RemoveContainer" containerID="0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8" Oct 09 10:23:46 crc kubenswrapper[4923]: E1009 10:23:46.208132 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8\": container with ID starting with 0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8 not found: ID does not exist" containerID="0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.208178 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8"} err="failed to get container status \"0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8\": rpc error: code = NotFound desc = could not find container \"0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8\": container with ID starting with 0c3d82b891b10bd9c44f468860f24cbda245f419da94b6ae5ecc9b7d906985d8 not found: ID does not exist" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.208210 4923 scope.go:117] "RemoveContainer" containerID="7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad" Oct 09 10:23:46 crc kubenswrapper[4923]: E1009 10:23:46.208509 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad\": container with ID starting with 7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad not found: ID does not exist" containerID="7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.208554 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad"} err="failed to get container status \"7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad\": rpc error: code = NotFound desc = could not find container \"7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad\": container with ID starting with 7ec6b920e73da4b949e1a2594478e07a8f8ad59084e701938b516776ced5d0ad not found: ID does not exist" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.233656 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4d125b1-feb1-43f8-94ad-417f4e163822" (UID: "e4d125b1-feb1-43f8-94ad-417f4e163822"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.255058 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data" (OuterVolumeSpecName: "config-data") pod "e4d125b1-feb1-43f8-94ad-417f4e163822" (UID: "e4d125b1-feb1-43f8-94ad-417f4e163822"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.292151 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4d125b1-feb1-43f8-94ad-417f4e163822-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.292200 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjzzt\" (UniqueName: \"kubernetes.io/projected/e4d125b1-feb1-43f8-94ad-417f4e163822-kube-api-access-zjzzt\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.292217 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.292229 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.292243 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d125b1-feb1-43f8-94ad-417f4e163822-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.486057 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6744bb4d8-9hkdw"] Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.516274 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6744bb4d8-9hkdw"] Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.612355 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" path="/var/lib/kubelet/pods/e4d125b1-feb1-43f8-94ad-417f4e163822/volumes" Oct 09 10:23:46 crc kubenswrapper[4923]: I1009 10:23:46.771326 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 09 10:23:47 crc kubenswrapper[4923]: E1009 10:23:47.543311 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf250b2a7_fb69_4c40_b186_238f74472a44.slice/crio-3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a\": RecentStats: unable to find data in memory cache]" Oct 09 10:23:47 crc kubenswrapper[4923]: I1009 10:23:47.844201 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 09 10:23:48 crc kubenswrapper[4923]: I1009 10:23:48.931821 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:48 crc kubenswrapper[4923]: I1009 10:23:48.936075 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-84bc56c5fd-mzbzg" Oct 09 10:23:49 crc kubenswrapper[4923]: I1009 10:23:49.214726 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5c7cc6d588-zpbfk" Oct 09 10:23:52 crc kubenswrapper[4923]: I1009 10:23:52.079254 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-855475c586-p22m4" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.141:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.141:8443: connect: connection refused" Oct 09 10:23:52 crc kubenswrapper[4923]: I1009 10:23:52.079890 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-855475c586-p22m4" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.058169 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 09 10:23:53 crc kubenswrapper[4923]: E1009 10:23:53.059157 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api-log" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.059182 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api-log" Oct 09 10:23:53 crc kubenswrapper[4923]: E1009 10:23:53.059201 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.059209 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.059498 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.059535 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4d125b1-feb1-43f8-94ad-417f4e163822" containerName="barbican-api-log" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.060379 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.062822 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.063040 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-7z8d5" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.063395 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.066028 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.110899 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.148748 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn82b\" (UniqueName: \"kubernetes.io/projected/9746f4ee-2c4e-4f93-9860-7536efb42c93-kube-api-access-bn82b\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.148927 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.149050 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.149130 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config-secret\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.250728 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config-secret\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.250876 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn82b\" (UniqueName: \"kubernetes.io/projected/9746f4ee-2c4e-4f93-9860-7536efb42c93-kube-api-access-bn82b\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.250952 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.251035 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.253291 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.258566 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config-secret\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.261429 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.271569 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn82b\" (UniqueName: \"kubernetes.io/projected/9746f4ee-2c4e-4f93-9860-7536efb42c93-kube-api-access-bn82b\") pod \"openstackclient\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.390782 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.397818 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.407233 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.436445 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.453340 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.453482 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: E1009 10:23:53.536260 4923 log.go:32] "RunPodSandbox from runtime service failed" err=< Oct 09 10:23:53 crc kubenswrapper[4923]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_9746f4ee-2c4e-4f93-9860-7536efb42c93_0(845657bc44dd5a518a4d543bf7efcd80fbd7c867351cbbeceeebdbf77cee3569): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"845657bc44dd5a518a4d543bf7efcd80fbd7c867351cbbeceeebdbf77cee3569" Netns:"/var/run/netns/cd00de83-fa3e-444c-bd4e-7e7f6fbf49c5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=845657bc44dd5a518a4d543bf7efcd80fbd7c867351cbbeceeebdbf77cee3569;K8S_POD_UID=9746f4ee-2c4e-4f93-9860-7536efb42c93" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/9746f4ee-2c4e-4f93-9860-7536efb42c93]: expected pod UID "9746f4ee-2c4e-4f93-9860-7536efb42c93" but got "bb06dc17-1098-420f-9106-4ea25ae8bd9d" from Kube API Oct 09 10:23:53 crc kubenswrapper[4923]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 09 10:23:53 crc kubenswrapper[4923]: > Oct 09 10:23:53 crc kubenswrapper[4923]: E1009 10:23:53.536362 4923 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Oct 09 10:23:53 crc kubenswrapper[4923]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_9746f4ee-2c4e-4f93-9860-7536efb42c93_0(845657bc44dd5a518a4d543bf7efcd80fbd7c867351cbbeceeebdbf77cee3569): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"845657bc44dd5a518a4d543bf7efcd80fbd7c867351cbbeceeebdbf77cee3569" Netns:"/var/run/netns/cd00de83-fa3e-444c-bd4e-7e7f6fbf49c5" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=845657bc44dd5a518a4d543bf7efcd80fbd7c867351cbbeceeebdbf77cee3569;K8S_POD_UID=9746f4ee-2c4e-4f93-9860-7536efb42c93" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/9746f4ee-2c4e-4f93-9860-7536efb42c93]: expected pod UID "9746f4ee-2c4e-4f93-9860-7536efb42c93" but got "bb06dc17-1098-420f-9106-4ea25ae8bd9d" from Kube API Oct 09 10:23:53 crc kubenswrapper[4923]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Oct 09 10:23:53 crc kubenswrapper[4923]: > pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.556886 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb06dc17-1098-420f-9106-4ea25ae8bd9d-openstack-config\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.557005 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgg7h\" (UniqueName: \"kubernetes.io/projected/bb06dc17-1098-420f-9106-4ea25ae8bd9d-kube-api-access-lgg7h\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.557545 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb06dc17-1098-420f-9106-4ea25ae8bd9d-openstack-config-secret\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.557682 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb06dc17-1098-420f-9106-4ea25ae8bd9d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.660371 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgg7h\" (UniqueName: \"kubernetes.io/projected/bb06dc17-1098-420f-9106-4ea25ae8bd9d-kube-api-access-lgg7h\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.660480 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb06dc17-1098-420f-9106-4ea25ae8bd9d-openstack-config-secret\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.660519 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb06dc17-1098-420f-9106-4ea25ae8bd9d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.660635 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb06dc17-1098-420f-9106-4ea25ae8bd9d-openstack-config\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.661642 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/bb06dc17-1098-420f-9106-4ea25ae8bd9d-openstack-config\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.668860 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/bb06dc17-1098-420f-9106-4ea25ae8bd9d-openstack-config-secret\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.669003 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb06dc17-1098-420f-9106-4ea25ae8bd9d-combined-ca-bundle\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.681299 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgg7h\" (UniqueName: \"kubernetes.io/projected/bb06dc17-1098-420f-9106-4ea25ae8bd9d-kube-api-access-lgg7h\") pod \"openstackclient\" (UID: \"bb06dc17-1098-420f-9106-4ea25ae8bd9d\") " pod="openstack/openstackclient" Oct 09 10:23:53 crc kubenswrapper[4923]: I1009 10:23:53.885209 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.241918 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.246912 4923 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9746f4ee-2c4e-4f93-9860-7536efb42c93" podUID="bb06dc17-1098-420f-9106-4ea25ae8bd9d" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.260332 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.374793 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config\") pod \"9746f4ee-2c4e-4f93-9860-7536efb42c93\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.374986 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config-secret\") pod \"9746f4ee-2c4e-4f93-9860-7536efb42c93\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.375128 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-combined-ca-bundle\") pod \"9746f4ee-2c4e-4f93-9860-7536efb42c93\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.375176 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bn82b\" (UniqueName: \"kubernetes.io/projected/9746f4ee-2c4e-4f93-9860-7536efb42c93-kube-api-access-bn82b\") pod \"9746f4ee-2c4e-4f93-9860-7536efb42c93\" (UID: \"9746f4ee-2c4e-4f93-9860-7536efb42c93\") " Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.375485 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "9746f4ee-2c4e-4f93-9860-7536efb42c93" (UID: "9746f4ee-2c4e-4f93-9860-7536efb42c93"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.376390 4923 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.379337 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.382439 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "9746f4ee-2c4e-4f93-9860-7536efb42c93" (UID: "9746f4ee-2c4e-4f93-9860-7536efb42c93"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.382628 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9746f4ee-2c4e-4f93-9860-7536efb42c93-kube-api-access-bn82b" (OuterVolumeSpecName: "kube-api-access-bn82b") pod "9746f4ee-2c4e-4f93-9860-7536efb42c93" (UID: "9746f4ee-2c4e-4f93-9860-7536efb42c93"). InnerVolumeSpecName "kube-api-access-bn82b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.385776 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9746f4ee-2c4e-4f93-9860-7536efb42c93" (UID: "9746f4ee-2c4e-4f93-9860-7536efb42c93"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.478114 4923 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.478581 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9746f4ee-2c4e-4f93-9860-7536efb42c93-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.478596 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bn82b\" (UniqueName: \"kubernetes.io/projected/9746f4ee-2c4e-4f93-9860-7536efb42c93-kube-api-access-bn82b\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.600150 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.600243 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:23:54 crc kubenswrapper[4923]: I1009 10:23:54.613427 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9746f4ee-2c4e-4f93-9860-7536efb42c93" path="/var/lib/kubelet/pods/9746f4ee-2c4e-4f93-9860-7536efb42c93/volumes" Oct 09 10:23:55 crc kubenswrapper[4923]: I1009 10:23:55.253197 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 09 10:23:55 crc kubenswrapper[4923]: I1009 10:23:55.253399 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bb06dc17-1098-420f-9106-4ea25ae8bd9d","Type":"ContainerStarted","Data":"d302ab2d93b18fe74a5cd02475f991d9c6cfd5bbf0a5c81a81e254ce141fae0d"} Oct 09 10:23:55 crc kubenswrapper[4923]: I1009 10:23:55.260638 4923 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="9746f4ee-2c4e-4f93-9860-7536efb42c93" podUID="bb06dc17-1098-420f-9106-4ea25ae8bd9d" Oct 09 10:23:57 crc kubenswrapper[4923]: E1009 10:23:57.866983 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf250b2a7_fb69_4c40_b186_238f74472a44.slice/crio-3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f2b211c_36ed_4fe4_9c2d_4b1a3353c29b.slice/crio-conmon-fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2.scope\": RecentStats: unable to find data in memory cache]" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.174374 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-855475c586-p22m4" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.263020 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-secret-key\") pod \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.263106 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgw7x\" (UniqueName: \"kubernetes.io/projected/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-kube-api-access-cgw7x\") pod \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.263319 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-config-data\") pod \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.263346 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-tls-certs\") pod \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.263394 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-scripts\") pod \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.263414 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-combined-ca-bundle\") pod \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.263534 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-logs\") pod \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\" (UID: \"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b\") " Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.264340 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-logs" (OuterVolumeSpecName: "logs") pod "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" (UID: "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.270700 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" (UID: "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.270944 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-kube-api-access-cgw7x" (OuterVolumeSpecName: "kube-api-access-cgw7x") pod "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" (UID: "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b"). InnerVolumeSpecName "kube-api-access-cgw7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.290736 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-scripts" (OuterVolumeSpecName: "scripts") pod "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" (UID: "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.299457 4923 generic.go:334] "Generic (PLEG): container finished" podID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerID="fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2" exitCode=137 Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.299519 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855475c586-p22m4" event={"ID":"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b","Type":"ContainerDied","Data":"fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2"} Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.299565 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-855475c586-p22m4" event={"ID":"3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b","Type":"ContainerDied","Data":"afc3d7870ca910ccc67dce336fedeb9a97c4eba98f58587805294702f948d453"} Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.299590 4923 scope.go:117] "RemoveContainer" containerID="b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.299746 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-855475c586-p22m4" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.306102 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-config-data" (OuterVolumeSpecName: "config-data") pod "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" (UID: "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.315130 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" (UID: "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.332895 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" (UID: "3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.366231 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.366274 4923 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.366289 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgw7x\" (UniqueName: \"kubernetes.io/projected/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-kube-api-access-cgw7x\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.366316 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.366328 4923 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.366339 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.366350 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.535892 4923 scope.go:117] "RemoveContainer" containerID="fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2" Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.640985 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-855475c586-p22m4"] Oct 09 10:23:58 crc kubenswrapper[4923]: I1009 10:23:58.651424 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-855475c586-p22m4"] Oct 09 10:24:00 crc kubenswrapper[4923]: I1009 10:24:00.614376 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" path="/var/lib/kubelet/pods/3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b/volumes" Oct 09 10:24:01 crc kubenswrapper[4923]: I1009 10:24:01.825843 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-nwwwr"] Oct 09 10:24:01 crc kubenswrapper[4923]: E1009 10:24:01.830265 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon" Oct 09 10:24:01 crc kubenswrapper[4923]: I1009 10:24:01.830325 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon" Oct 09 10:24:01 crc kubenswrapper[4923]: E1009 10:24:01.830355 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon-log" Oct 09 10:24:01 crc kubenswrapper[4923]: I1009 10:24:01.830365 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon-log" Oct 09 10:24:01 crc kubenswrapper[4923]: I1009 10:24:01.861419 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon" Oct 09 10:24:01 crc kubenswrapper[4923]: I1009 10:24:01.861537 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2b211c-36ed-4fe4-9c2d-4b1a3353c29b" containerName="horizon-log" Oct 09 10:24:01 crc kubenswrapper[4923]: I1009 10:24:01.862570 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nwwwr"] Oct 09 10:24:01 crc kubenswrapper[4923]: I1009 10:24:01.862704 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nwwwr" Oct 09 10:24:01 crc kubenswrapper[4923]: I1009 10:24:01.951898 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g2hg\" (UniqueName: \"kubernetes.io/projected/667c88d0-d5c3-416c-b166-b68f127ce657-kube-api-access-8g2hg\") pod \"nova-api-db-create-nwwwr\" (UID: \"667c88d0-d5c3-416c-b166-b68f127ce657\") " pod="openstack/nova-api-db-create-nwwwr" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.011960 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-ckntf"] Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.013551 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ckntf" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.032986 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ckntf"] Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.054731 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g2hg\" (UniqueName: \"kubernetes.io/projected/667c88d0-d5c3-416c-b166-b68f127ce657-kube-api-access-8g2hg\") pod \"nova-api-db-create-nwwwr\" (UID: \"667c88d0-d5c3-416c-b166-b68f127ce657\") " pod="openstack/nova-api-db-create-nwwwr" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.098708 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g2hg\" (UniqueName: \"kubernetes.io/projected/667c88d0-d5c3-416c-b166-b68f127ce657-kube-api-access-8g2hg\") pod \"nova-api-db-create-nwwwr\" (UID: \"667c88d0-d5c3-416c-b166-b68f127ce657\") " pod="openstack/nova-api-db-create-nwwwr" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.118691 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-7pgnz"] Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.120084 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7pgnz" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.130262 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7pgnz"] Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.158307 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxmtp\" (UniqueName: \"kubernetes.io/projected/4747efe7-6523-4c92-82d3-ce5d2a8b2868-kube-api-access-rxmtp\") pod \"nova-cell0-db-create-ckntf\" (UID: \"4747efe7-6523-4c92-82d3-ce5d2a8b2868\") " pod="openstack/nova-cell0-db-create-ckntf" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.216502 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nwwwr" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.260132 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnrz5\" (UniqueName: \"kubernetes.io/projected/a32a19c5-4480-4d2b-9563-208e6a0ced35-kube-api-access-lnrz5\") pod \"nova-cell1-db-create-7pgnz\" (UID: \"a32a19c5-4480-4d2b-9563-208e6a0ced35\") " pod="openstack/nova-cell1-db-create-7pgnz" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.260563 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxmtp\" (UniqueName: \"kubernetes.io/projected/4747efe7-6523-4c92-82d3-ce5d2a8b2868-kube-api-access-rxmtp\") pod \"nova-cell0-db-create-ckntf\" (UID: \"4747efe7-6523-4c92-82d3-ce5d2a8b2868\") " pod="openstack/nova-cell0-db-create-ckntf" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.292401 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxmtp\" (UniqueName: \"kubernetes.io/projected/4747efe7-6523-4c92-82d3-ce5d2a8b2868-kube-api-access-rxmtp\") pod \"nova-cell0-db-create-ckntf\" (UID: \"4747efe7-6523-4c92-82d3-ce5d2a8b2868\") " pod="openstack/nova-cell0-db-create-ckntf" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.330963 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ckntf" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.362902 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnrz5\" (UniqueName: \"kubernetes.io/projected/a32a19c5-4480-4d2b-9563-208e6a0ced35-kube-api-access-lnrz5\") pod \"nova-cell1-db-create-7pgnz\" (UID: \"a32a19c5-4480-4d2b-9563-208e6a0ced35\") " pod="openstack/nova-cell1-db-create-7pgnz" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.397360 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnrz5\" (UniqueName: \"kubernetes.io/projected/a32a19c5-4480-4d2b-9563-208e6a0ced35-kube-api-access-lnrz5\") pod \"nova-cell1-db-create-7pgnz\" (UID: \"a32a19c5-4480-4d2b-9563-208e6a0ced35\") " pod="openstack/nova-cell1-db-create-7pgnz" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.480312 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7pgnz" Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.872689 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.873539 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="ceilometer-central-agent" containerID="cri-o://c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369" gracePeriod=30 Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.874078 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="proxy-httpd" containerID="cri-o://0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb" gracePeriod=30 Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.874136 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="sg-core" containerID="cri-o://ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0" gracePeriod=30 Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.874172 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="ceilometer-notification-agent" containerID="cri-o://39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec" gracePeriod=30 Oct 09 10:24:02 crc kubenswrapper[4923]: I1009 10:24:02.881281 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 09 10:24:03 crc kubenswrapper[4923]: I1009 10:24:03.368709 4923 generic.go:334] "Generic (PLEG): container finished" podID="5598fa65-0ab7-4995-b47c-d144192572bc" containerID="0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb" exitCode=0 Oct 09 10:24:03 crc kubenswrapper[4923]: I1009 10:24:03.368830 4923 generic.go:334] "Generic (PLEG): container finished" podID="5598fa65-0ab7-4995-b47c-d144192572bc" containerID="ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0" exitCode=2 Oct 09 10:24:03 crc kubenswrapper[4923]: I1009 10:24:03.368829 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerDied","Data":"0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb"} Oct 09 10:24:03 crc kubenswrapper[4923]: I1009 10:24:03.368881 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerDied","Data":"ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0"} Oct 09 10:24:03 crc kubenswrapper[4923]: I1009 10:24:03.368895 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerDied","Data":"c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369"} Oct 09 10:24:03 crc kubenswrapper[4923]: I1009 10:24:03.368847 4923 generic.go:334] "Generic (PLEG): container finished" podID="5598fa65-0ab7-4995-b47c-d144192572bc" containerID="c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369" exitCode=0 Oct 09 10:24:04 crc kubenswrapper[4923]: I1009 10:24:04.657409 4923 scope.go:117] "RemoveContainer" containerID="b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667" Oct 09 10:24:04 crc kubenswrapper[4923]: E1009 10:24:04.662218 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667\": container with ID starting with b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667 not found: ID does not exist" containerID="b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667" Oct 09 10:24:04 crc kubenswrapper[4923]: I1009 10:24:04.662266 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667"} err="failed to get container status \"b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667\": rpc error: code = NotFound desc = could not find container \"b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667\": container with ID starting with b800a5388b9991714dab39064d23394f10c8398c3329f36c8b31b199c2b41667 not found: ID does not exist" Oct 09 10:24:04 crc kubenswrapper[4923]: I1009 10:24:04.662294 4923 scope.go:117] "RemoveContainer" containerID="fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2" Oct 09 10:24:04 crc kubenswrapper[4923]: E1009 10:24:04.662676 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2\": container with ID starting with fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2 not found: ID does not exist" containerID="fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2" Oct 09 10:24:04 crc kubenswrapper[4923]: I1009 10:24:04.662771 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2"} err="failed to get container status \"fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2\": rpc error: code = NotFound desc = could not find container \"fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2\": container with ID starting with fe03ec3c2033e7035aa47638e4960a1a0c4c8ba035e34200425260388ea1cee2 not found: ID does not exist" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.087152 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.226560 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-run-httpd\") pod \"5598fa65-0ab7-4995-b47c-d144192572bc\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.227315 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxstj\" (UniqueName: \"kubernetes.io/projected/5598fa65-0ab7-4995-b47c-d144192572bc-kube-api-access-qxstj\") pod \"5598fa65-0ab7-4995-b47c-d144192572bc\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.227366 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-sg-core-conf-yaml\") pod \"5598fa65-0ab7-4995-b47c-d144192572bc\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.227042 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5598fa65-0ab7-4995-b47c-d144192572bc" (UID: "5598fa65-0ab7-4995-b47c-d144192572bc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.227413 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-config-data\") pod \"5598fa65-0ab7-4995-b47c-d144192572bc\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.227559 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-combined-ca-bundle\") pod \"5598fa65-0ab7-4995-b47c-d144192572bc\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.227648 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-log-httpd\") pod \"5598fa65-0ab7-4995-b47c-d144192572bc\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.227680 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-scripts\") pod \"5598fa65-0ab7-4995-b47c-d144192572bc\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.228823 4923 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.233229 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5598fa65-0ab7-4995-b47c-d144192572bc" (UID: "5598fa65-0ab7-4995-b47c-d144192572bc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.249283 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-scripts" (OuterVolumeSpecName: "scripts") pod "5598fa65-0ab7-4995-b47c-d144192572bc" (UID: "5598fa65-0ab7-4995-b47c-d144192572bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.260387 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5598fa65-0ab7-4995-b47c-d144192572bc-kube-api-access-qxstj" (OuterVolumeSpecName: "kube-api-access-qxstj") pod "5598fa65-0ab7-4995-b47c-d144192572bc" (UID: "5598fa65-0ab7-4995-b47c-d144192572bc"). InnerVolumeSpecName "kube-api-access-qxstj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.276002 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5598fa65-0ab7-4995-b47c-d144192572bc" (UID: "5598fa65-0ab7-4995-b47c-d144192572bc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.330510 4923 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5598fa65-0ab7-4995-b47c-d144192572bc-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.330575 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.330586 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxstj\" (UniqueName: \"kubernetes.io/projected/5598fa65-0ab7-4995-b47c-d144192572bc-kube-api-access-qxstj\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.330599 4923 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.349630 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-ckntf"] Oct 09 10:24:05 crc kubenswrapper[4923]: W1009 10:24:05.359867 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda32a19c5_4480_4d2b_9563_208e6a0ced35.slice/crio-82e0d1cf6517b06ea8beb44349036b086d404cd570ad46747a52f4b40dac925f WatchSource:0}: Error finding container 82e0d1cf6517b06ea8beb44349036b086d404cd570ad46747a52f4b40dac925f: Status 404 returned error can't find the container with id 82e0d1cf6517b06ea8beb44349036b086d404cd570ad46747a52f4b40dac925f Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.387832 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5598fa65-0ab7-4995-b47c-d144192572bc" (UID: "5598fa65-0ab7-4995-b47c-d144192572bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.403179 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-nwwwr"] Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.411737 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"bb06dc17-1098-420f-9106-4ea25ae8bd9d","Type":"ContainerStarted","Data":"d7b9fc0eee0ffcc7808b6802e0cf399014597475ff634df315838d668084708d"} Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.423497 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7pgnz" event={"ID":"a32a19c5-4480-4d2b-9563-208e6a0ced35","Type":"ContainerStarted","Data":"82e0d1cf6517b06ea8beb44349036b086d404cd570ad46747a52f4b40dac925f"} Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.424983 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ckntf" event={"ID":"4747efe7-6523-4c92-82d3-ce5d2a8b2868","Type":"ContainerStarted","Data":"aab5b356d27b059ec604b3bca76a6d07c31b8e507d754d72dd8e15dae56de11d"} Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.429097 4923 generic.go:334] "Generic (PLEG): container finished" podID="5598fa65-0ab7-4995-b47c-d144192572bc" containerID="39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec" exitCode=0 Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.429182 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerDied","Data":"39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec"} Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.429222 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5598fa65-0ab7-4995-b47c-d144192572bc","Type":"ContainerDied","Data":"5daba828d09fa5281e4eac51650ad1cbd95d5910b1d39ab9e900c8d9571eecb9"} Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.429247 4923 scope.go:117] "RemoveContainer" containerID="0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.429458 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.432161 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-7pgnz"] Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.432421 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-config-data" (OuterVolumeSpecName: "config-data") pod "5598fa65-0ab7-4995-b47c-d144192572bc" (UID: "5598fa65-0ab7-4995-b47c-d144192572bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.432613 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-config-data\") pod \"5598fa65-0ab7-4995-b47c-d144192572bc\" (UID: \"5598fa65-0ab7-4995-b47c-d144192572bc\") " Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.433138 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:05 crc kubenswrapper[4923]: W1009 10:24:05.433239 4923 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/5598fa65-0ab7-4995-b47c-d144192572bc/volumes/kubernetes.io~secret/config-data Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.433252 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-config-data" (OuterVolumeSpecName: "config-data") pod "5598fa65-0ab7-4995-b47c-d144192572bc" (UID: "5598fa65-0ab7-4995-b47c-d144192572bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.441365 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nwwwr" event={"ID":"667c88d0-d5c3-416c-b166-b68f127ce657","Type":"ContainerStarted","Data":"e714f42768ada8ac0f330a05d0be0a34c853156a39ca3584dc61f8850ca842b2"} Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.485999 4923 scope.go:117] "RemoveContainer" containerID="ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.535191 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5598fa65-0ab7-4995-b47c-d144192572bc-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.551638 4923 scope.go:117] "RemoveContainer" containerID="39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.660886 4923 scope.go:117] "RemoveContainer" containerID="c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.720770 4923 scope.go:117] "RemoveContainer" containerID="0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb" Oct 09 10:24:05 crc kubenswrapper[4923]: E1009 10:24:05.721858 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb\": container with ID starting with 0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb not found: ID does not exist" containerID="0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.722457 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb"} err="failed to get container status \"0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb\": rpc error: code = NotFound desc = could not find container \"0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb\": container with ID starting with 0eb404effe9348fcfb29da9b5cc478597afea4a8a0f5a72d3e69a4790e1704cb not found: ID does not exist" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.722502 4923 scope.go:117] "RemoveContainer" containerID="ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0" Oct 09 10:24:05 crc kubenswrapper[4923]: E1009 10:24:05.723124 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0\": container with ID starting with ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0 not found: ID does not exist" containerID="ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.723182 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0"} err="failed to get container status \"ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0\": rpc error: code = NotFound desc = could not find container \"ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0\": container with ID starting with ec27b2d56ebfe4cd50b4bf1db1fec1c1867f39897b8bf622e5fad7ee15d453b0 not found: ID does not exist" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.723220 4923 scope.go:117] "RemoveContainer" containerID="39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec" Oct 09 10:24:05 crc kubenswrapper[4923]: E1009 10:24:05.723731 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec\": container with ID starting with 39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec not found: ID does not exist" containerID="39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.723778 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec"} err="failed to get container status \"39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec\": rpc error: code = NotFound desc = could not find container \"39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec\": container with ID starting with 39550bf8c48c4036e0abb2cdfb3bacc2f74754f9932609bd2ea494afe83cabec not found: ID does not exist" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.723798 4923 scope.go:117] "RemoveContainer" containerID="c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369" Oct 09 10:24:05 crc kubenswrapper[4923]: E1009 10:24:05.724187 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369\": container with ID starting with c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369 not found: ID does not exist" containerID="c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.724251 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369"} err="failed to get container status \"c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369\": rpc error: code = NotFound desc = could not find container \"c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369\": container with ID starting with c306238e0ea72aceba1a351b439bc430bc2daf37ed2ca53945ac5838b3684369 not found: ID does not exist" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.780777 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.498779021 podStartE2EDuration="12.780730484s" podCreationTimestamp="2025-10-09 10:23:53 +0000 UTC" firstStartedPulling="2025-10-09 10:23:54.383205828 +0000 UTC m=+1120.451387584" lastFinishedPulling="2025-10-09 10:24:04.665157291 +0000 UTC m=+1130.733339047" observedRunningTime="2025-10-09 10:24:05.457939721 +0000 UTC m=+1131.526121477" watchObservedRunningTime="2025-10-09 10:24:05.780730484 +0000 UTC m=+1131.848912240" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.789353 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.799188 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.822038 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:05 crc kubenswrapper[4923]: E1009 10:24:05.822503 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="ceilometer-notification-agent" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.822527 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="ceilometer-notification-agent" Oct 09 10:24:05 crc kubenswrapper[4923]: E1009 10:24:05.822538 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="ceilometer-central-agent" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.822544 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="ceilometer-central-agent" Oct 09 10:24:05 crc kubenswrapper[4923]: E1009 10:24:05.822579 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="proxy-httpd" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.822587 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="proxy-httpd" Oct 09 10:24:05 crc kubenswrapper[4923]: E1009 10:24:05.822602 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="sg-core" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.822607 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="sg-core" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.822809 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="sg-core" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.822824 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="ceilometer-notification-agent" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.822837 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="proxy-httpd" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.822846 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" containerName="ceilometer-central-agent" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.824786 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.828067 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.828298 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.853081 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.943281 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-run-httpd\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.943366 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-config-data\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.943511 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-log-httpd\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.943550 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-scripts\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.943577 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.943625 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:05 crc kubenswrapper[4923]: I1009 10:24:05.943651 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8mfg\" (UniqueName: \"kubernetes.io/projected/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-kube-api-access-h8mfg\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.046046 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-run-httpd\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.046306 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-config-data\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.046382 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-log-httpd\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.046423 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-scripts\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.046450 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.046509 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.046538 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8mfg\" (UniqueName: \"kubernetes.io/projected/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-kube-api-access-h8mfg\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.046837 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-run-httpd\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.047847 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-log-httpd\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.055920 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.057458 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-config-data\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.057611 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.061684 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-scripts\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.068342 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8mfg\" (UniqueName: \"kubernetes.io/projected/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-kube-api-access-h8mfg\") pod \"ceilometer-0\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.144365 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.455916 4923 generic.go:334] "Generic (PLEG): container finished" podID="a32a19c5-4480-4d2b-9563-208e6a0ced35" containerID="1713cb05129fc81548f8522b7c9d09d32cc96f5a802add679b382b97fcb7743f" exitCode=0 Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.456019 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7pgnz" event={"ID":"a32a19c5-4480-4d2b-9563-208e6a0ced35","Type":"ContainerDied","Data":"1713cb05129fc81548f8522b7c9d09d32cc96f5a802add679b382b97fcb7743f"} Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.459044 4923 generic.go:334] "Generic (PLEG): container finished" podID="4747efe7-6523-4c92-82d3-ce5d2a8b2868" containerID="5cff3d46f1024cb5cf1326d24532df562280dd8442d8f50b896e203e5b64b080" exitCode=0 Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.459137 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ckntf" event={"ID":"4747efe7-6523-4c92-82d3-ce5d2a8b2868","Type":"ContainerDied","Data":"5cff3d46f1024cb5cf1326d24532df562280dd8442d8f50b896e203e5b64b080"} Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.461462 4923 generic.go:334] "Generic (PLEG): container finished" podID="667c88d0-d5c3-416c-b166-b68f127ce657" containerID="4a342738a782fcf429cbbe261eb1c984df2b63b7a692620ca8d41c949a0c09fb" exitCode=0 Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.462939 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nwwwr" event={"ID":"667c88d0-d5c3-416c-b166-b68f127ce657","Type":"ContainerDied","Data":"4a342738a782fcf429cbbe261eb1c984df2b63b7a692620ca8d41c949a0c09fb"} Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.614016 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5598fa65-0ab7-4995-b47c-d144192572bc" path="/var/lib/kubelet/pods/5598fa65-0ab7-4995-b47c-d144192572bc/volumes" Oct 09 10:24:06 crc kubenswrapper[4923]: I1009 10:24:06.641008 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:07 crc kubenswrapper[4923]: I1009 10:24:07.474937 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerStarted","Data":"8152d42b8778ea2ed19a05894d840bcfbd3b9bbe31ebce8e467131aad7ade460"} Oct 09 10:24:07 crc kubenswrapper[4923]: I1009 10:24:07.475372 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerStarted","Data":"6539e41a6addb5116c96dc4ff67d23feda7d6e520a4d72d5f8863987d1df8f83"} Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.080032 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7pgnz" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.087953 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ckntf" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.098712 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nwwwr" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.192464 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8g2hg\" (UniqueName: \"kubernetes.io/projected/667c88d0-d5c3-416c-b166-b68f127ce657-kube-api-access-8g2hg\") pod \"667c88d0-d5c3-416c-b166-b68f127ce657\" (UID: \"667c88d0-d5c3-416c-b166-b68f127ce657\") " Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.193030 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxmtp\" (UniqueName: \"kubernetes.io/projected/4747efe7-6523-4c92-82d3-ce5d2a8b2868-kube-api-access-rxmtp\") pod \"4747efe7-6523-4c92-82d3-ce5d2a8b2868\" (UID: \"4747efe7-6523-4c92-82d3-ce5d2a8b2868\") " Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.193078 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnrz5\" (UniqueName: \"kubernetes.io/projected/a32a19c5-4480-4d2b-9563-208e6a0ced35-kube-api-access-lnrz5\") pod \"a32a19c5-4480-4d2b-9563-208e6a0ced35\" (UID: \"a32a19c5-4480-4d2b-9563-208e6a0ced35\") " Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.199596 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a32a19c5-4480-4d2b-9563-208e6a0ced35-kube-api-access-lnrz5" (OuterVolumeSpecName: "kube-api-access-lnrz5") pod "a32a19c5-4480-4d2b-9563-208e6a0ced35" (UID: "a32a19c5-4480-4d2b-9563-208e6a0ced35"). InnerVolumeSpecName "kube-api-access-lnrz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.200373 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/667c88d0-d5c3-416c-b166-b68f127ce657-kube-api-access-8g2hg" (OuterVolumeSpecName: "kube-api-access-8g2hg") pod "667c88d0-d5c3-416c-b166-b68f127ce657" (UID: "667c88d0-d5c3-416c-b166-b68f127ce657"). InnerVolumeSpecName "kube-api-access-8g2hg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.203500 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4747efe7-6523-4c92-82d3-ce5d2a8b2868-kube-api-access-rxmtp" (OuterVolumeSpecName: "kube-api-access-rxmtp") pod "4747efe7-6523-4c92-82d3-ce5d2a8b2868" (UID: "4747efe7-6523-4c92-82d3-ce5d2a8b2868"). InnerVolumeSpecName "kube-api-access-rxmtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.295928 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxmtp\" (UniqueName: \"kubernetes.io/projected/4747efe7-6523-4c92-82d3-ce5d2a8b2868-kube-api-access-rxmtp\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.295975 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnrz5\" (UniqueName: \"kubernetes.io/projected/a32a19c5-4480-4d2b-9563-208e6a0ced35-kube-api-access-lnrz5\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.295988 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8g2hg\" (UniqueName: \"kubernetes.io/projected/667c88d0-d5c3-416c-b166-b68f127ce657-kube-api-access-8g2hg\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:08 crc kubenswrapper[4923]: E1009 10:24:08.375789 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf250b2a7_fb69_4c40_b186_238f74472a44.slice/crio-3131c0539e04f6bfa2f2f5af56fded38af38459b6728ecbc52b98493808e840a\": RecentStats: unable to find data in memory cache]" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.486935 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-ckntf" event={"ID":"4747efe7-6523-4c92-82d3-ce5d2a8b2868","Type":"ContainerDied","Data":"aab5b356d27b059ec604b3bca76a6d07c31b8e507d754d72dd8e15dae56de11d"} Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.487382 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aab5b356d27b059ec604b3bca76a6d07c31b8e507d754d72dd8e15dae56de11d" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.487041 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-ckntf" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.489380 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-nwwwr" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.489422 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-nwwwr" event={"ID":"667c88d0-d5c3-416c-b166-b68f127ce657","Type":"ContainerDied","Data":"e714f42768ada8ac0f330a05d0be0a34c853156a39ca3584dc61f8850ca842b2"} Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.489472 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e714f42768ada8ac0f330a05d0be0a34c853156a39ca3584dc61f8850ca842b2" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.491354 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-7pgnz" event={"ID":"a32a19c5-4480-4d2b-9563-208e6a0ced35","Type":"ContainerDied","Data":"82e0d1cf6517b06ea8beb44349036b086d404cd570ad46747a52f4b40dac925f"} Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.491394 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="82e0d1cf6517b06ea8beb44349036b086d404cd570ad46747a52f4b40dac925f" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.491389 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-7pgnz" Oct 09 10:24:08 crc kubenswrapper[4923]: I1009 10:24:08.494008 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerStarted","Data":"73de5ddfe7ca44f864f4d762a4da4f2d88306dd406c946f7a0d06f7c9b31b1c6"} Oct 09 10:24:09 crc kubenswrapper[4923]: I1009 10:24:09.082199 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:09 crc kubenswrapper[4923]: I1009 10:24:09.506125 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerStarted","Data":"a6c73d0b213fe6ca21fcede041aa6d2131eef7204254d1e982e405afaac8e47d"} Oct 09 10:24:10 crc kubenswrapper[4923]: I1009 10:24:10.531213 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerStarted","Data":"6a29cc2eba31666ca3eb97b9074527da561b54c7d22d68d4d748d136ea58a0a8"} Oct 09 10:24:10 crc kubenswrapper[4923]: I1009 10:24:10.531710 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 10:24:10 crc kubenswrapper[4923]: I1009 10:24:10.531479 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="sg-core" containerID="cri-o://a6c73d0b213fe6ca21fcede041aa6d2131eef7204254d1e982e405afaac8e47d" gracePeriod=30 Oct 09 10:24:10 crc kubenswrapper[4923]: I1009 10:24:10.531435 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="ceilometer-central-agent" containerID="cri-o://8152d42b8778ea2ed19a05894d840bcfbd3b9bbe31ebce8e467131aad7ade460" gracePeriod=30 Oct 09 10:24:10 crc kubenswrapper[4923]: I1009 10:24:10.531493 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="proxy-httpd" containerID="cri-o://6a29cc2eba31666ca3eb97b9074527da561b54c7d22d68d4d748d136ea58a0a8" gracePeriod=30 Oct 09 10:24:10 crc kubenswrapper[4923]: I1009 10:24:10.531499 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="ceilometer-notification-agent" containerID="cri-o://73de5ddfe7ca44f864f4d762a4da4f2d88306dd406c946f7a0d06f7c9b31b1c6" gracePeriod=30 Oct 09 10:24:10 crc kubenswrapper[4923]: I1009 10:24:10.561622 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.032140069 podStartE2EDuration="5.561601108s" podCreationTimestamp="2025-10-09 10:24:05 +0000 UTC" firstStartedPulling="2025-10-09 10:24:06.644659209 +0000 UTC m=+1132.712840965" lastFinishedPulling="2025-10-09 10:24:10.174120248 +0000 UTC m=+1136.242302004" observedRunningTime="2025-10-09 10:24:10.559294714 +0000 UTC m=+1136.627476480" watchObservedRunningTime="2025-10-09 10:24:10.561601108 +0000 UTC m=+1136.629782864" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.545403 4923 generic.go:334] "Generic (PLEG): container finished" podID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerID="6a29cc2eba31666ca3eb97b9074527da561b54c7d22d68d4d748d136ea58a0a8" exitCode=0 Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.545782 4923 generic.go:334] "Generic (PLEG): container finished" podID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerID="a6c73d0b213fe6ca21fcede041aa6d2131eef7204254d1e982e405afaac8e47d" exitCode=2 Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.545795 4923 generic.go:334] "Generic (PLEG): container finished" podID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerID="73de5ddfe7ca44f864f4d762a4da4f2d88306dd406c946f7a0d06f7c9b31b1c6" exitCode=0 Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.545479 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerDied","Data":"6a29cc2eba31666ca3eb97b9074527da561b54c7d22d68d4d748d136ea58a0a8"} Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.545837 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerDied","Data":"a6c73d0b213fe6ca21fcede041aa6d2131eef7204254d1e982e405afaac8e47d"} Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.545856 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerDied","Data":"73de5ddfe7ca44f864f4d762a4da4f2d88306dd406c946f7a0d06f7c9b31b1c6"} Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.972259 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-c481-account-create-mfwgm"] Oct 09 10:24:11 crc kubenswrapper[4923]: E1009 10:24:11.977252 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="667c88d0-d5c3-416c-b166-b68f127ce657" containerName="mariadb-database-create" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.977289 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="667c88d0-d5c3-416c-b166-b68f127ce657" containerName="mariadb-database-create" Oct 09 10:24:11 crc kubenswrapper[4923]: E1009 10:24:11.977315 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4747efe7-6523-4c92-82d3-ce5d2a8b2868" containerName="mariadb-database-create" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.977321 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="4747efe7-6523-4c92-82d3-ce5d2a8b2868" containerName="mariadb-database-create" Oct 09 10:24:11 crc kubenswrapper[4923]: E1009 10:24:11.977340 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a32a19c5-4480-4d2b-9563-208e6a0ced35" containerName="mariadb-database-create" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.977365 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="a32a19c5-4480-4d2b-9563-208e6a0ced35" containerName="mariadb-database-create" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.977590 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="a32a19c5-4480-4d2b-9563-208e6a0ced35" containerName="mariadb-database-create" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.977612 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="4747efe7-6523-4c92-82d3-ce5d2a8b2868" containerName="mariadb-database-create" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.977629 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="667c88d0-d5c3-416c-b166-b68f127ce657" containerName="mariadb-database-create" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.978447 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c481-account-create-mfwgm" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.981488 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 09 10:24:11 crc kubenswrapper[4923]: I1009 10:24:11.983067 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c481-account-create-mfwgm"] Oct 09 10:24:12 crc kubenswrapper[4923]: I1009 10:24:12.077369 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mg4l5\" (UniqueName: \"kubernetes.io/projected/6917f504-25e3-4991-a7cb-1f319586f9dd-kube-api-access-mg4l5\") pod \"nova-api-c481-account-create-mfwgm\" (UID: \"6917f504-25e3-4991-a7cb-1f319586f9dd\") " pod="openstack/nova-api-c481-account-create-mfwgm" Oct 09 10:24:12 crc kubenswrapper[4923]: I1009 10:24:12.179352 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mg4l5\" (UniqueName: \"kubernetes.io/projected/6917f504-25e3-4991-a7cb-1f319586f9dd-kube-api-access-mg4l5\") pod \"nova-api-c481-account-create-mfwgm\" (UID: \"6917f504-25e3-4991-a7cb-1f319586f9dd\") " pod="openstack/nova-api-c481-account-create-mfwgm" Oct 09 10:24:12 crc kubenswrapper[4923]: I1009 10:24:12.206979 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mg4l5\" (UniqueName: \"kubernetes.io/projected/6917f504-25e3-4991-a7cb-1f319586f9dd-kube-api-access-mg4l5\") pod \"nova-api-c481-account-create-mfwgm\" (UID: \"6917f504-25e3-4991-a7cb-1f319586f9dd\") " pod="openstack/nova-api-c481-account-create-mfwgm" Oct 09 10:24:12 crc kubenswrapper[4923]: I1009 10:24:12.295377 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c481-account-create-mfwgm" Oct 09 10:24:12 crc kubenswrapper[4923]: I1009 10:24:12.758741 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-c481-account-create-mfwgm"] Oct 09 10:24:13 crc kubenswrapper[4923]: I1009 10:24:13.581645 4923 generic.go:334] "Generic (PLEG): container finished" podID="6917f504-25e3-4991-a7cb-1f319586f9dd" containerID="d3b3c2bbaec1c007eb581c20d7746c9adfd9bb779dcf06b99200d616a0a922f6" exitCode=0 Oct 09 10:24:13 crc kubenswrapper[4923]: I1009 10:24:13.581797 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c481-account-create-mfwgm" event={"ID":"6917f504-25e3-4991-a7cb-1f319586f9dd","Type":"ContainerDied","Data":"d3b3c2bbaec1c007eb581c20d7746c9adfd9bb779dcf06b99200d616a0a922f6"} Oct 09 10:24:13 crc kubenswrapper[4923]: I1009 10:24:13.582053 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c481-account-create-mfwgm" event={"ID":"6917f504-25e3-4991-a7cb-1f319586f9dd","Type":"ContainerStarted","Data":"285932ea42eb606531a340fd3b6cbe44d2a0ffa9c561c42e30a676d9f3a45bc4"} Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.628430 4923 generic.go:334] "Generic (PLEG): container finished" podID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerID="8152d42b8778ea2ed19a05894d840bcfbd3b9bbe31ebce8e467131aad7ade460" exitCode=0 Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.644895 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerDied","Data":"8152d42b8778ea2ed19a05894d840bcfbd3b9bbe31ebce8e467131aad7ade460"} Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.786120 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.849048 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8mfg\" (UniqueName: \"kubernetes.io/projected/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-kube-api-access-h8mfg\") pod \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.849122 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-sg-core-conf-yaml\") pod \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.849213 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-config-data\") pod \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.849374 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-run-httpd\") pod \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.849498 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-scripts\") pod \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.849590 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-combined-ca-bundle\") pod \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.849624 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-log-httpd\") pod \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\" (UID: \"d88358f5-af63-4e5b-8dbc-3314bd25d0b7\") " Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.850363 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d88358f5-af63-4e5b-8dbc-3314bd25d0b7" (UID: "d88358f5-af63-4e5b-8dbc-3314bd25d0b7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.850615 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d88358f5-af63-4e5b-8dbc-3314bd25d0b7" (UID: "d88358f5-af63-4e5b-8dbc-3314bd25d0b7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.860011 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-scripts" (OuterVolumeSpecName: "scripts") pod "d88358f5-af63-4e5b-8dbc-3314bd25d0b7" (UID: "d88358f5-af63-4e5b-8dbc-3314bd25d0b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.867193 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-kube-api-access-h8mfg" (OuterVolumeSpecName: "kube-api-access-h8mfg") pod "d88358f5-af63-4e5b-8dbc-3314bd25d0b7" (UID: "d88358f5-af63-4e5b-8dbc-3314bd25d0b7"). InnerVolumeSpecName "kube-api-access-h8mfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.922651 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d88358f5-af63-4e5b-8dbc-3314bd25d0b7" (UID: "d88358f5-af63-4e5b-8dbc-3314bd25d0b7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.952553 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.953477 4923 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.953555 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8mfg\" (UniqueName: \"kubernetes.io/projected/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-kube-api-access-h8mfg\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.953689 4923 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.953802 4923 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.977446 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d88358f5-af63-4e5b-8dbc-3314bd25d0b7" (UID: "d88358f5-af63-4e5b-8dbc-3314bd25d0b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.983907 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c481-account-create-mfwgm" Oct 09 10:24:14 crc kubenswrapper[4923]: I1009 10:24:14.988503 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-config-data" (OuterVolumeSpecName: "config-data") pod "d88358f5-af63-4e5b-8dbc-3314bd25d0b7" (UID: "d88358f5-af63-4e5b-8dbc-3314bd25d0b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.054938 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg4l5\" (UniqueName: \"kubernetes.io/projected/6917f504-25e3-4991-a7cb-1f319586f9dd-kube-api-access-mg4l5\") pod \"6917f504-25e3-4991-a7cb-1f319586f9dd\" (UID: \"6917f504-25e3-4991-a7cb-1f319586f9dd\") " Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.055569 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.055595 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d88358f5-af63-4e5b-8dbc-3314bd25d0b7-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.059293 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6917f504-25e3-4991-a7cb-1f319586f9dd-kube-api-access-mg4l5" (OuterVolumeSpecName: "kube-api-access-mg4l5") pod "6917f504-25e3-4991-a7cb-1f319586f9dd" (UID: "6917f504-25e3-4991-a7cb-1f319586f9dd"). InnerVolumeSpecName "kube-api-access-mg4l5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.158284 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg4l5\" (UniqueName: \"kubernetes.io/projected/6917f504-25e3-4991-a7cb-1f319586f9dd-kube-api-access-mg4l5\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.642594 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d88358f5-af63-4e5b-8dbc-3314bd25d0b7","Type":"ContainerDied","Data":"6539e41a6addb5116c96dc4ff67d23feda7d6e520a4d72d5f8863987d1df8f83"} Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.642641 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.642693 4923 scope.go:117] "RemoveContainer" containerID="6a29cc2eba31666ca3eb97b9074527da561b54c7d22d68d4d748d136ea58a0a8" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.647198 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-c481-account-create-mfwgm" event={"ID":"6917f504-25e3-4991-a7cb-1f319586f9dd","Type":"ContainerDied","Data":"285932ea42eb606531a340fd3b6cbe44d2a0ffa9c561c42e30a676d9f3a45bc4"} Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.647250 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="285932ea42eb606531a340fd3b6cbe44d2a0ffa9c561c42e30a676d9f3a45bc4" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.647279 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-c481-account-create-mfwgm" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.665722 4923 scope.go:117] "RemoveContainer" containerID="a6c73d0b213fe6ca21fcede041aa6d2131eef7204254d1e982e405afaac8e47d" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.693800 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.703998 4923 scope.go:117] "RemoveContainer" containerID="73de5ddfe7ca44f864f4d762a4da4f2d88306dd406c946f7a0d06f7c9b31b1c6" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.710743 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.733512 4923 scope.go:117] "RemoveContainer" containerID="8152d42b8778ea2ed19a05894d840bcfbd3b9bbe31ebce8e467131aad7ade460" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.736657 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:15 crc kubenswrapper[4923]: E1009 10:24:15.737280 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6917f504-25e3-4991-a7cb-1f319586f9dd" containerName="mariadb-account-create" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.737304 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="6917f504-25e3-4991-a7cb-1f319586f9dd" containerName="mariadb-account-create" Oct 09 10:24:15 crc kubenswrapper[4923]: E1009 10:24:15.737323 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="ceilometer-central-agent" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.737351 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="ceilometer-central-agent" Oct 09 10:24:15 crc kubenswrapper[4923]: E1009 10:24:15.737379 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="ceilometer-notification-agent" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.737385 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="ceilometer-notification-agent" Oct 09 10:24:15 crc kubenswrapper[4923]: E1009 10:24:15.738637 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="proxy-httpd" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.738653 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="proxy-httpd" Oct 09 10:24:15 crc kubenswrapper[4923]: E1009 10:24:15.738677 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="sg-core" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.738685 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="sg-core" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.738968 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="ceilometer-central-agent" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.738991 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="sg-core" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.739008 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="6917f504-25e3-4991-a7cb-1f319586f9dd" containerName="mariadb-account-create" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.739028 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="ceilometer-notification-agent" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.739048 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" containerName="proxy-httpd" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.743399 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.747592 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.748569 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.750554 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.876058 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-log-httpd\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.876136 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-run-httpd\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.876505 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.876576 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-config-data\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.876681 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.876943 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-scripts\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.877146 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwqgc\" (UniqueName: \"kubernetes.io/projected/fd48ee6b-6602-4480-9035-9bcb920fad7b-kube-api-access-pwqgc\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.978600 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.978640 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-config-data\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.978675 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.978709 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-scripts\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.978834 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwqgc\" (UniqueName: \"kubernetes.io/projected/fd48ee6b-6602-4480-9035-9bcb920fad7b-kube-api-access-pwqgc\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.978869 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-log-httpd\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.978901 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-run-httpd\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.979504 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-run-httpd\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.981066 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-log-httpd\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.987793 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.988628 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-config-data\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:15 crc kubenswrapper[4923]: I1009 10:24:15.995286 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-scripts\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:16 crc kubenswrapper[4923]: I1009 10:24:16.006718 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwqgc\" (UniqueName: \"kubernetes.io/projected/fd48ee6b-6602-4480-9035-9bcb920fad7b-kube-api-access-pwqgc\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:16 crc kubenswrapper[4923]: I1009 10:24:16.006816 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " pod="openstack/ceilometer-0" Oct 09 10:24:16 crc kubenswrapper[4923]: I1009 10:24:16.070227 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:16 crc kubenswrapper[4923]: I1009 10:24:16.597847 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:16 crc kubenswrapper[4923]: I1009 10:24:16.613429 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d88358f5-af63-4e5b-8dbc-3314bd25d0b7" path="/var/lib/kubelet/pods/d88358f5-af63-4e5b-8dbc-3314bd25d0b7/volumes" Oct 09 10:24:16 crc kubenswrapper[4923]: I1009 10:24:16.659895 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerStarted","Data":"3eaed9b05a74ab4625e4a3d30636f527f3793c53dc361ea2e1a6abbe105f8d32"} Oct 09 10:24:16 crc kubenswrapper[4923]: I1009 10:24:16.915129 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:17 crc kubenswrapper[4923]: I1009 10:24:17.670739 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerStarted","Data":"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188"} Oct 09 10:24:18 crc kubenswrapper[4923]: I1009 10:24:18.684769 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerStarted","Data":"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904"} Oct 09 10:24:19 crc kubenswrapper[4923]: I1009 10:24:19.695197 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerStarted","Data":"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e"} Oct 09 10:24:21 crc kubenswrapper[4923]: I1009 10:24:21.728143 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerStarted","Data":"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57"} Oct 09 10:24:21 crc kubenswrapper[4923]: I1009 10:24:21.728790 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="ceilometer-central-agent" containerID="cri-o://a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188" gracePeriod=30 Oct 09 10:24:21 crc kubenswrapper[4923]: I1009 10:24:21.729167 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 10:24:21 crc kubenswrapper[4923]: I1009 10:24:21.729572 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="proxy-httpd" containerID="cri-o://f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57" gracePeriod=30 Oct 09 10:24:21 crc kubenswrapper[4923]: I1009 10:24:21.729650 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="sg-core" containerID="cri-o://07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e" gracePeriod=30 Oct 09 10:24:21 crc kubenswrapper[4923]: I1009 10:24:21.729741 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="ceilometer-notification-agent" containerID="cri-o://b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904" gracePeriod=30 Oct 09 10:24:21 crc kubenswrapper[4923]: I1009 10:24:21.769363 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.949063325 podStartE2EDuration="6.769329858s" podCreationTimestamp="2025-10-09 10:24:15 +0000 UTC" firstStartedPulling="2025-10-09 10:24:16.604937071 +0000 UTC m=+1142.673118827" lastFinishedPulling="2025-10-09 10:24:20.425203604 +0000 UTC m=+1146.493385360" observedRunningTime="2025-10-09 10:24:21.760951258 +0000 UTC m=+1147.829133044" watchObservedRunningTime="2025-10-09 10:24:21.769329858 +0000 UTC m=+1147.837511624" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.189609 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-b76b-account-create-27zj7"] Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.191467 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b76b-account-create-27zj7" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.193938 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.214516 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b76b-account-create-27zj7"] Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.333820 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdlpc\" (UniqueName: \"kubernetes.io/projected/aeadce81-f642-4762-be0f-20650018847f-kube-api-access-sdlpc\") pod \"nova-cell0-b76b-account-create-27zj7\" (UID: \"aeadce81-f642-4762-be0f-20650018847f\") " pod="openstack/nova-cell0-b76b-account-create-27zj7" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.389392 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-ec7c-account-create-7ppmh"] Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.393340 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ec7c-account-create-7ppmh" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.397603 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.401563 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ec7c-account-create-7ppmh"] Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.435905 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdlpc\" (UniqueName: \"kubernetes.io/projected/aeadce81-f642-4762-be0f-20650018847f-kube-api-access-sdlpc\") pod \"nova-cell0-b76b-account-create-27zj7\" (UID: \"aeadce81-f642-4762-be0f-20650018847f\") " pod="openstack/nova-cell0-b76b-account-create-27zj7" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.459579 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdlpc\" (UniqueName: \"kubernetes.io/projected/aeadce81-f642-4762-be0f-20650018847f-kube-api-access-sdlpc\") pod \"nova-cell0-b76b-account-create-27zj7\" (UID: \"aeadce81-f642-4762-be0f-20650018847f\") " pod="openstack/nova-cell0-b76b-account-create-27zj7" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.538514 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztmsz\" (UniqueName: \"kubernetes.io/projected/9637ebcc-b8d8-4f61-ae1b-372802dfa9ec-kube-api-access-ztmsz\") pod \"nova-cell1-ec7c-account-create-7ppmh\" (UID: \"9637ebcc-b8d8-4f61-ae1b-372802dfa9ec\") " pod="openstack/nova-cell1-ec7c-account-create-7ppmh" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.549738 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.623244 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b76b-account-create-27zj7" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.640365 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwqgc\" (UniqueName: \"kubernetes.io/projected/fd48ee6b-6602-4480-9035-9bcb920fad7b-kube-api-access-pwqgc\") pod \"fd48ee6b-6602-4480-9035-9bcb920fad7b\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.641345 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-sg-core-conf-yaml\") pod \"fd48ee6b-6602-4480-9035-9bcb920fad7b\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.642344 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-config-data\") pod \"fd48ee6b-6602-4480-9035-9bcb920fad7b\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.642416 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-run-httpd\") pod \"fd48ee6b-6602-4480-9035-9bcb920fad7b\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.642476 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-combined-ca-bundle\") pod \"fd48ee6b-6602-4480-9035-9bcb920fad7b\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.642519 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-log-httpd\") pod \"fd48ee6b-6602-4480-9035-9bcb920fad7b\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.642703 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-scripts\") pod \"fd48ee6b-6602-4480-9035-9bcb920fad7b\" (UID: \"fd48ee6b-6602-4480-9035-9bcb920fad7b\") " Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.642903 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fd48ee6b-6602-4480-9035-9bcb920fad7b" (UID: "fd48ee6b-6602-4480-9035-9bcb920fad7b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.643162 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fd48ee6b-6602-4480-9035-9bcb920fad7b" (UID: "fd48ee6b-6602-4480-9035-9bcb920fad7b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.644769 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztmsz\" (UniqueName: \"kubernetes.io/projected/9637ebcc-b8d8-4f61-ae1b-372802dfa9ec-kube-api-access-ztmsz\") pod \"nova-cell1-ec7c-account-create-7ppmh\" (UID: \"9637ebcc-b8d8-4f61-ae1b-372802dfa9ec\") " pod="openstack/nova-cell1-ec7c-account-create-7ppmh" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.644942 4923 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.644956 4923 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fd48ee6b-6602-4480-9035-9bcb920fad7b-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.649673 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-scripts" (OuterVolumeSpecName: "scripts") pod "fd48ee6b-6602-4480-9035-9bcb920fad7b" (UID: "fd48ee6b-6602-4480-9035-9bcb920fad7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.649768 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd48ee6b-6602-4480-9035-9bcb920fad7b-kube-api-access-pwqgc" (OuterVolumeSpecName: "kube-api-access-pwqgc") pod "fd48ee6b-6602-4480-9035-9bcb920fad7b" (UID: "fd48ee6b-6602-4480-9035-9bcb920fad7b"). InnerVolumeSpecName "kube-api-access-pwqgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.668108 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztmsz\" (UniqueName: \"kubernetes.io/projected/9637ebcc-b8d8-4f61-ae1b-372802dfa9ec-kube-api-access-ztmsz\") pod \"nova-cell1-ec7c-account-create-7ppmh\" (UID: \"9637ebcc-b8d8-4f61-ae1b-372802dfa9ec\") " pod="openstack/nova-cell1-ec7c-account-create-7ppmh" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.675510 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fd48ee6b-6602-4480-9035-9bcb920fad7b" (UID: "fd48ee6b-6602-4480-9035-9bcb920fad7b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.710674 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ec7c-account-create-7ppmh" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.747689 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.747719 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwqgc\" (UniqueName: \"kubernetes.io/projected/fd48ee6b-6602-4480-9035-9bcb920fad7b-kube-api-access-pwqgc\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.747730 4923 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.762847 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd48ee6b-6602-4480-9035-9bcb920fad7b" (UID: "fd48ee6b-6602-4480-9035-9bcb920fad7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.767119 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-config-data" (OuterVolumeSpecName: "config-data") pod "fd48ee6b-6602-4480-9035-9bcb920fad7b" (UID: "fd48ee6b-6602-4480-9035-9bcb920fad7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.771932 4923 generic.go:334] "Generic (PLEG): container finished" podID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerID="f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57" exitCode=0 Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.771962 4923 generic.go:334] "Generic (PLEG): container finished" podID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerID="07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e" exitCode=2 Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.771973 4923 generic.go:334] "Generic (PLEG): container finished" podID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerID="b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904" exitCode=0 Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.771984 4923 generic.go:334] "Generic (PLEG): container finished" podID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerID="a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188" exitCode=0 Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.772013 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerDied","Data":"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57"} Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.772058 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerDied","Data":"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e"} Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.772073 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerDied","Data":"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904"} Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.772085 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerDied","Data":"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188"} Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.772095 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fd48ee6b-6602-4480-9035-9bcb920fad7b","Type":"ContainerDied","Data":"3eaed9b05a74ab4625e4a3d30636f527f3793c53dc361ea2e1a6abbe105f8d32"} Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.772115 4923 scope.go:117] "RemoveContainer" containerID="f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.772380 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.810457 4923 scope.go:117] "RemoveContainer" containerID="07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.834907 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.883620 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.884519 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.884547 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd48ee6b-6602-4480-9035-9bcb920fad7b-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.920984 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:22 crc kubenswrapper[4923]: E1009 10:24:22.921583 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="ceilometer-central-agent" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.921603 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="ceilometer-central-agent" Oct 09 10:24:22 crc kubenswrapper[4923]: E1009 10:24:22.921624 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="proxy-httpd" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.921632 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="proxy-httpd" Oct 09 10:24:22 crc kubenswrapper[4923]: E1009 10:24:22.921653 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="ceilometer-notification-agent" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.921662 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="ceilometer-notification-agent" Oct 09 10:24:22 crc kubenswrapper[4923]: E1009 10:24:22.921684 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="sg-core" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.921691 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="sg-core" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.921937 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="ceilometer-central-agent" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.921966 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="proxy-httpd" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.921990 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="ceilometer-notification-agent" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.922001 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" containerName="sg-core" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.925332 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.929981 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.930024 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.930472 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.935546 4923 scope.go:117] "RemoveContainer" containerID="b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.964153 4923 scope.go:117] "RemoveContainer" containerID="a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.995280 4923 scope.go:117] "RemoveContainer" containerID="f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57" Oct 09 10:24:22 crc kubenswrapper[4923]: E1009 10:24:22.995997 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57\": container with ID starting with f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57 not found: ID does not exist" containerID="f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.996036 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57"} err="failed to get container status \"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57\": rpc error: code = NotFound desc = could not find container \"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57\": container with ID starting with f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57 not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.996071 4923 scope.go:117] "RemoveContainer" containerID="07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e" Oct 09 10:24:22 crc kubenswrapper[4923]: E1009 10:24:22.996485 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e\": container with ID starting with 07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e not found: ID does not exist" containerID="07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.996553 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e"} err="failed to get container status \"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e\": rpc error: code = NotFound desc = could not find container \"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e\": container with ID starting with 07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.996598 4923 scope.go:117] "RemoveContainer" containerID="b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904" Oct 09 10:24:22 crc kubenswrapper[4923]: E1009 10:24:22.996997 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904\": container with ID starting with b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904 not found: ID does not exist" containerID="b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.997036 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904"} err="failed to get container status \"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904\": rpc error: code = NotFound desc = could not find container \"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904\": container with ID starting with b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904 not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.997058 4923 scope.go:117] "RemoveContainer" containerID="a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188" Oct 09 10:24:22 crc kubenswrapper[4923]: E1009 10:24:22.997608 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188\": container with ID starting with a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188 not found: ID does not exist" containerID="a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.997650 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188"} err="failed to get container status \"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188\": rpc error: code = NotFound desc = could not find container \"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188\": container with ID starting with a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188 not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.997669 4923 scope.go:117] "RemoveContainer" containerID="f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.998103 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57"} err="failed to get container status \"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57\": rpc error: code = NotFound desc = could not find container \"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57\": container with ID starting with f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57 not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.998134 4923 scope.go:117] "RemoveContainer" containerID="07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.998440 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e"} err="failed to get container status \"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e\": rpc error: code = NotFound desc = could not find container \"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e\": container with ID starting with 07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.998518 4923 scope.go:117] "RemoveContainer" containerID="b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.998809 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904"} err="failed to get container status \"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904\": rpc error: code = NotFound desc = could not find container \"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904\": container with ID starting with b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904 not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.998841 4923 scope.go:117] "RemoveContainer" containerID="a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.999160 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188"} err="failed to get container status \"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188\": rpc error: code = NotFound desc = could not find container \"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188\": container with ID starting with a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188 not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.999184 4923 scope.go:117] "RemoveContainer" containerID="f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.999455 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57"} err="failed to get container status \"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57\": rpc error: code = NotFound desc = could not find container \"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57\": container with ID starting with f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57 not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.999478 4923 scope.go:117] "RemoveContainer" containerID="07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.999799 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e"} err="failed to get container status \"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e\": rpc error: code = NotFound desc = could not find container \"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e\": container with ID starting with 07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e not found: ID does not exist" Oct 09 10:24:22 crc kubenswrapper[4923]: I1009 10:24:22.999835 4923 scope.go:117] "RemoveContainer" containerID="b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.000131 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904"} err="failed to get container status \"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904\": rpc error: code = NotFound desc = could not find container \"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904\": container with ID starting with b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904 not found: ID does not exist" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.000160 4923 scope.go:117] "RemoveContainer" containerID="a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.000499 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188"} err="failed to get container status \"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188\": rpc error: code = NotFound desc = could not find container \"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188\": container with ID starting with a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188 not found: ID does not exist" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.000521 4923 scope.go:117] "RemoveContainer" containerID="f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.001031 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57"} err="failed to get container status \"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57\": rpc error: code = NotFound desc = could not find container \"f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57\": container with ID starting with f7ef370654ad50590397f9fc2f95ce984c3330627ad0e80b786a03dfd386eb57 not found: ID does not exist" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.001061 4923 scope.go:117] "RemoveContainer" containerID="07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.001505 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e"} err="failed to get container status \"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e\": rpc error: code = NotFound desc = could not find container \"07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e\": container with ID starting with 07cd5f9a98287cd1c7a341e367284e8b154e594d6e62de31f8408be4ba683e3e not found: ID does not exist" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.001566 4923 scope.go:117] "RemoveContainer" containerID="b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.004539 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904"} err="failed to get container status \"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904\": rpc error: code = NotFound desc = could not find container \"b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904\": container with ID starting with b10db8ce6cdda406f06d1e133e12d09e78aaca480b6bbd7a989e0d1181bac904 not found: ID does not exist" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.004567 4923 scope.go:117] "RemoveContainer" containerID="a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.005057 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188"} err="failed to get container status \"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188\": rpc error: code = NotFound desc = could not find container \"a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188\": container with ID starting with a945bc7821c01c9bfe137a227dad85290f0c4f3fbb4384c8b1e50281a32b7188 not found: ID does not exist" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.088856 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-run-httpd\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.088982 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.089034 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-config-data\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.089056 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.089078 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-log-httpd\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.089316 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75gqp\" (UniqueName: \"kubernetes.io/projected/e6bd9476-77c4-45af-a38d-6972483c61c1-kube-api-access-75gqp\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.089544 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-scripts\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.150210 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-b76b-account-create-27zj7"] Oct 09 10:24:23 crc kubenswrapper[4923]: W1009 10:24:23.166288 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaeadce81_f642_4762_be0f_20650018847f.slice/crio-df302dbedded2f849de5173913445384e81fb338b884c77f369eace0714f01f5 WatchSource:0}: Error finding container df302dbedded2f849de5173913445384e81fb338b884c77f369eace0714f01f5: Status 404 returned error can't find the container with id df302dbedded2f849de5173913445384e81fb338b884c77f369eace0714f01f5 Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.191899 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75gqp\" (UniqueName: \"kubernetes.io/projected/e6bd9476-77c4-45af-a38d-6972483c61c1-kube-api-access-75gqp\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.192042 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-scripts\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.192577 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-run-httpd\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.192665 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.192789 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-config-data\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.192815 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.192864 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-log-httpd\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.193386 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-run-httpd\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.193420 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-log-httpd\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.198825 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.199416 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.204581 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-scripts\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.206550 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-config-data\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.213518 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75gqp\" (UniqueName: \"kubernetes.io/projected/e6bd9476-77c4-45af-a38d-6972483c61c1-kube-api-access-75gqp\") pod \"ceilometer-0\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.255836 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.285200 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-ec7c-account-create-7ppmh"] Oct 09 10:24:23 crc kubenswrapper[4923]: W1009 10:24:23.299524 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9637ebcc_b8d8_4f61_ae1b_372802dfa9ec.slice/crio-430c84dc8456def6a712bcf25677557f537be2bdde51dd24bd3899fdd93e66f8 WatchSource:0}: Error finding container 430c84dc8456def6a712bcf25677557f537be2bdde51dd24bd3899fdd93e66f8: Status 404 returned error can't find the container with id 430c84dc8456def6a712bcf25677557f537be2bdde51dd24bd3899fdd93e66f8 Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.755678 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:23 crc kubenswrapper[4923]: W1009 10:24:23.761393 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode6bd9476_77c4_45af_a38d_6972483c61c1.slice/crio-b2fa7110582e9356adc4c0b6657b9b437ed4aa3b101c8b141713b33b97577915 WatchSource:0}: Error finding container b2fa7110582e9356adc4c0b6657b9b437ed4aa3b101c8b141713b33b97577915: Status 404 returned error can't find the container with id b2fa7110582e9356adc4c0b6657b9b437ed4aa3b101c8b141713b33b97577915 Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.789817 4923 generic.go:334] "Generic (PLEG): container finished" podID="aeadce81-f642-4762-be0f-20650018847f" containerID="2771c19824fd85f6e780d7e3e52832c155a101f65e3ba2b0d6e39419acfb2575" exitCode=0 Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.789936 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b76b-account-create-27zj7" event={"ID":"aeadce81-f642-4762-be0f-20650018847f","Type":"ContainerDied","Data":"2771c19824fd85f6e780d7e3e52832c155a101f65e3ba2b0d6e39419acfb2575"} Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.790032 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b76b-account-create-27zj7" event={"ID":"aeadce81-f642-4762-be0f-20650018847f","Type":"ContainerStarted","Data":"df302dbedded2f849de5173913445384e81fb338b884c77f369eace0714f01f5"} Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.795485 4923 generic.go:334] "Generic (PLEG): container finished" podID="9637ebcc-b8d8-4f61-ae1b-372802dfa9ec" containerID="a0a8a6793f398684ad8ecdf254131a51037766c132b042b977951beae89080f1" exitCode=0 Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.795575 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ec7c-account-create-7ppmh" event={"ID":"9637ebcc-b8d8-4f61-ae1b-372802dfa9ec","Type":"ContainerDied","Data":"a0a8a6793f398684ad8ecdf254131a51037766c132b042b977951beae89080f1"} Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.795621 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ec7c-account-create-7ppmh" event={"ID":"9637ebcc-b8d8-4f61-ae1b-372802dfa9ec","Type":"ContainerStarted","Data":"430c84dc8456def6a712bcf25677557f537be2bdde51dd24bd3899fdd93e66f8"} Oct 09 10:24:23 crc kubenswrapper[4923]: I1009 10:24:23.797250 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerStarted","Data":"b2fa7110582e9356adc4c0b6657b9b437ed4aa3b101c8b141713b33b97577915"} Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.600093 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.600514 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.612711 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd48ee6b-6602-4480-9035-9bcb920fad7b" path="/var/lib/kubelet/pods/fd48ee6b-6602-4480-9035-9bcb920fad7b/volumes" Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.613877 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.614438 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ff1594437e0fe053768cc52266cdd8b6d71b1a9aae0c236740cf019156ef7f8f"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.614523 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://ff1594437e0fe053768cc52266cdd8b6d71b1a9aae0c236740cf019156ef7f8f" gracePeriod=600 Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.808793 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerStarted","Data":"d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3"} Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.811874 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="ff1594437e0fe053768cc52266cdd8b6d71b1a9aae0c236740cf019156ef7f8f" exitCode=0 Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.811966 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"ff1594437e0fe053768cc52266cdd8b6d71b1a9aae0c236740cf019156ef7f8f"} Oct 09 10:24:24 crc kubenswrapper[4923]: I1009 10:24:24.812085 4923 scope.go:117] "RemoveContainer" containerID="0371acf1b4e04ebc9cc4694cbecac34256a8ea0829820a4e57e504509cef2cbf" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.208379 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b76b-account-create-27zj7" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.215951 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ec7c-account-create-7ppmh" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.348961 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdlpc\" (UniqueName: \"kubernetes.io/projected/aeadce81-f642-4762-be0f-20650018847f-kube-api-access-sdlpc\") pod \"aeadce81-f642-4762-be0f-20650018847f\" (UID: \"aeadce81-f642-4762-be0f-20650018847f\") " Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.349100 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztmsz\" (UniqueName: \"kubernetes.io/projected/9637ebcc-b8d8-4f61-ae1b-372802dfa9ec-kube-api-access-ztmsz\") pod \"9637ebcc-b8d8-4f61-ae1b-372802dfa9ec\" (UID: \"9637ebcc-b8d8-4f61-ae1b-372802dfa9ec\") " Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.357844 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aeadce81-f642-4762-be0f-20650018847f-kube-api-access-sdlpc" (OuterVolumeSpecName: "kube-api-access-sdlpc") pod "aeadce81-f642-4762-be0f-20650018847f" (UID: "aeadce81-f642-4762-be0f-20650018847f"). InnerVolumeSpecName "kube-api-access-sdlpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.357924 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9637ebcc-b8d8-4f61-ae1b-372802dfa9ec-kube-api-access-ztmsz" (OuterVolumeSpecName: "kube-api-access-ztmsz") pod "9637ebcc-b8d8-4f61-ae1b-372802dfa9ec" (UID: "9637ebcc-b8d8-4f61-ae1b-372802dfa9ec"). InnerVolumeSpecName "kube-api-access-ztmsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.452335 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztmsz\" (UniqueName: \"kubernetes.io/projected/9637ebcc-b8d8-4f61-ae1b-372802dfa9ec-kube-api-access-ztmsz\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.452381 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdlpc\" (UniqueName: \"kubernetes.io/projected/aeadce81-f642-4762-be0f-20650018847f-kube-api-access-sdlpc\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.825791 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"cb8062901f893bbba9ad244080627bdc62960c7f3035d2ee9483ebb1559dcc72"} Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.829328 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-b76b-account-create-27zj7" event={"ID":"aeadce81-f642-4762-be0f-20650018847f","Type":"ContainerDied","Data":"df302dbedded2f849de5173913445384e81fb338b884c77f369eace0714f01f5"} Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.829392 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df302dbedded2f849de5173913445384e81fb338b884c77f369eace0714f01f5" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.829346 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-b76b-account-create-27zj7" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.831420 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-ec7c-account-create-7ppmh" event={"ID":"9637ebcc-b8d8-4f61-ae1b-372802dfa9ec","Type":"ContainerDied","Data":"430c84dc8456def6a712bcf25677557f537be2bdde51dd24bd3899fdd93e66f8"} Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.831423 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-ec7c-account-create-7ppmh" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.831464 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="430c84dc8456def6a712bcf25677557f537be2bdde51dd24bd3899fdd93e66f8" Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.841376 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerStarted","Data":"3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372"} Oct 09 10:24:25 crc kubenswrapper[4923]: I1009 10:24:25.841441 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerStarted","Data":"dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6"} Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.422164 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x5qkv"] Oct 09 10:24:27 crc kubenswrapper[4923]: E1009 10:24:27.423389 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aeadce81-f642-4762-be0f-20650018847f" containerName="mariadb-account-create" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.423405 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="aeadce81-f642-4762-be0f-20650018847f" containerName="mariadb-account-create" Oct 09 10:24:27 crc kubenswrapper[4923]: E1009 10:24:27.423450 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9637ebcc-b8d8-4f61-ae1b-372802dfa9ec" containerName="mariadb-account-create" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.423459 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9637ebcc-b8d8-4f61-ae1b-372802dfa9ec" containerName="mariadb-account-create" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.423688 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="aeadce81-f642-4762-be0f-20650018847f" containerName="mariadb-account-create" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.423720 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9637ebcc-b8d8-4f61-ae1b-372802dfa9ec" containerName="mariadb-account-create" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.424607 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.427126 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.427386 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.427550 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-jnl25" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.437064 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x5qkv"] Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.603257 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.603384 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l66cc\" (UniqueName: \"kubernetes.io/projected/31875d1d-5bd0-4a30-b82e-0113ad3cf237-kube-api-access-l66cc\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.603452 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-scripts\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.603658 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-config-data\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.705999 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-config-data\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.706624 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.706806 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l66cc\" (UniqueName: \"kubernetes.io/projected/31875d1d-5bd0-4a30-b82e-0113ad3cf237-kube-api-access-l66cc\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.706959 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-scripts\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.714113 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-scripts\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.715240 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-config-data\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.716287 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.724610 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l66cc\" (UniqueName: \"kubernetes.io/projected/31875d1d-5bd0-4a30-b82e-0113ad3cf237-kube-api-access-l66cc\") pod \"nova-cell0-conductor-db-sync-x5qkv\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.762813 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.866092 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerStarted","Data":"f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922"} Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.866432 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 10:24:27 crc kubenswrapper[4923]: I1009 10:24:27.901707 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.578669545 podStartE2EDuration="5.901676039s" podCreationTimestamp="2025-10-09 10:24:22 +0000 UTC" firstStartedPulling="2025-10-09 10:24:23.764242449 +0000 UTC m=+1149.832424225" lastFinishedPulling="2025-10-09 10:24:27.087248963 +0000 UTC m=+1153.155430719" observedRunningTime="2025-10-09 10:24:27.901291008 +0000 UTC m=+1153.969472764" watchObservedRunningTime="2025-10-09 10:24:27.901676039 +0000 UTC m=+1153.969857795" Oct 09 10:24:28 crc kubenswrapper[4923]: I1009 10:24:28.243711 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x5qkv"] Oct 09 10:24:28 crc kubenswrapper[4923]: I1009 10:24:28.921803 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x5qkv" event={"ID":"31875d1d-5bd0-4a30-b82e-0113ad3cf237","Type":"ContainerStarted","Data":"917f2b7427f0eafd34b8a33695ca039b0e12f40be471434f0030e6c2bccb0fe9"} Oct 09 10:24:36 crc kubenswrapper[4923]: I1009 10:24:36.016114 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x5qkv" event={"ID":"31875d1d-5bd0-4a30-b82e-0113ad3cf237","Type":"ContainerStarted","Data":"51f2e52140e71d0bc17a819b33beaca75a330900540d0853813b9400663ba51e"} Oct 09 10:24:36 crc kubenswrapper[4923]: I1009 10:24:36.045246 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-x5qkv" podStartSLOduration=1.777958054 podStartE2EDuration="9.045208888s" podCreationTimestamp="2025-10-09 10:24:27 +0000 UTC" firstStartedPulling="2025-10-09 10:24:28.267386761 +0000 UTC m=+1154.335568517" lastFinishedPulling="2025-10-09 10:24:35.534637605 +0000 UTC m=+1161.602819351" observedRunningTime="2025-10-09 10:24:36.03510768 +0000 UTC m=+1162.103289436" watchObservedRunningTime="2025-10-09 10:24:36.045208888 +0000 UTC m=+1162.113390664" Oct 09 10:24:52 crc kubenswrapper[4923]: I1009 10:24:52.183201 4923 generic.go:334] "Generic (PLEG): container finished" podID="31875d1d-5bd0-4a30-b82e-0113ad3cf237" containerID="51f2e52140e71d0bc17a819b33beaca75a330900540d0853813b9400663ba51e" exitCode=0 Oct 09 10:24:52 crc kubenswrapper[4923]: I1009 10:24:52.183300 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x5qkv" event={"ID":"31875d1d-5bd0-4a30-b82e-0113ad3cf237","Type":"ContainerDied","Data":"51f2e52140e71d0bc17a819b33beaca75a330900540d0853813b9400663ba51e"} Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.261837 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.589944 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.659524 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-combined-ca-bundle\") pod \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.660062 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l66cc\" (UniqueName: \"kubernetes.io/projected/31875d1d-5bd0-4a30-b82e-0113ad3cf237-kube-api-access-l66cc\") pod \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.660136 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-config-data\") pod \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.660240 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-scripts\") pod \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\" (UID: \"31875d1d-5bd0-4a30-b82e-0113ad3cf237\") " Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.667970 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31875d1d-5bd0-4a30-b82e-0113ad3cf237-kube-api-access-l66cc" (OuterVolumeSpecName: "kube-api-access-l66cc") pod "31875d1d-5bd0-4a30-b82e-0113ad3cf237" (UID: "31875d1d-5bd0-4a30-b82e-0113ad3cf237"). InnerVolumeSpecName "kube-api-access-l66cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.668414 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-scripts" (OuterVolumeSpecName: "scripts") pod "31875d1d-5bd0-4a30-b82e-0113ad3cf237" (UID: "31875d1d-5bd0-4a30-b82e-0113ad3cf237"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.694585 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-config-data" (OuterVolumeSpecName: "config-data") pod "31875d1d-5bd0-4a30-b82e-0113ad3cf237" (UID: "31875d1d-5bd0-4a30-b82e-0113ad3cf237"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.697921 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31875d1d-5bd0-4a30-b82e-0113ad3cf237" (UID: "31875d1d-5bd0-4a30-b82e-0113ad3cf237"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.764815 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.765052 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l66cc\" (UniqueName: \"kubernetes.io/projected/31875d1d-5bd0-4a30-b82e-0113ad3cf237-kube-api-access-l66cc\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.765075 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:53 crc kubenswrapper[4923]: I1009 10:24:53.765089 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31875d1d-5bd0-4a30-b82e-0113ad3cf237-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.205204 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-x5qkv" event={"ID":"31875d1d-5bd0-4a30-b82e-0113ad3cf237","Type":"ContainerDied","Data":"917f2b7427f0eafd34b8a33695ca039b0e12f40be471434f0030e6c2bccb0fe9"} Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.205263 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="917f2b7427f0eafd34b8a33695ca039b0e12f40be471434f0030e6c2bccb0fe9" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.205280 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-x5qkv" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.319534 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 09 10:24:54 crc kubenswrapper[4923]: E1009 10:24:54.320090 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31875d1d-5bd0-4a30-b82e-0113ad3cf237" containerName="nova-cell0-conductor-db-sync" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.320107 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="31875d1d-5bd0-4a30-b82e-0113ad3cf237" containerName="nova-cell0-conductor-db-sync" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.320312 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="31875d1d-5bd0-4a30-b82e-0113ad3cf237" containerName="nova-cell0-conductor-db-sync" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.321327 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.325996 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-jnl25" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.326888 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.338937 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.480049 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f52a73f-7675-4cd5-ae86-77bd49333940-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3f52a73f-7675-4cd5-ae86-77bd49333940\") " pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.480253 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f52a73f-7675-4cd5-ae86-77bd49333940-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3f52a73f-7675-4cd5-ae86-77bd49333940\") " pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.480356 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq2f9\" (UniqueName: \"kubernetes.io/projected/3f52a73f-7675-4cd5-ae86-77bd49333940-kube-api-access-jq2f9\") pod \"nova-cell0-conductor-0\" (UID: \"3f52a73f-7675-4cd5-ae86-77bd49333940\") " pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.582689 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f52a73f-7675-4cd5-ae86-77bd49333940-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3f52a73f-7675-4cd5-ae86-77bd49333940\") " pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.582866 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f52a73f-7675-4cd5-ae86-77bd49333940-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3f52a73f-7675-4cd5-ae86-77bd49333940\") " pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.582922 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq2f9\" (UniqueName: \"kubernetes.io/projected/3f52a73f-7675-4cd5-ae86-77bd49333940-kube-api-access-jq2f9\") pod \"nova-cell0-conductor-0\" (UID: \"3f52a73f-7675-4cd5-ae86-77bd49333940\") " pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.588565 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f52a73f-7675-4cd5-ae86-77bd49333940-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"3f52a73f-7675-4cd5-ae86-77bd49333940\") " pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.591154 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f52a73f-7675-4cd5-ae86-77bd49333940-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"3f52a73f-7675-4cd5-ae86-77bd49333940\") " pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.599789 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq2f9\" (UniqueName: \"kubernetes.io/projected/3f52a73f-7675-4cd5-ae86-77bd49333940-kube-api-access-jq2f9\") pod \"nova-cell0-conductor-0\" (UID: \"3f52a73f-7675-4cd5-ae86-77bd49333940\") " pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:54 crc kubenswrapper[4923]: I1009 10:24:54.647217 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:55 crc kubenswrapper[4923]: I1009 10:24:55.132421 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 09 10:24:55 crc kubenswrapper[4923]: I1009 10:24:55.220728 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3f52a73f-7675-4cd5-ae86-77bd49333940","Type":"ContainerStarted","Data":"2b9abef2d05fdf2a5e3c6d4454e20e735891acfd5aad26a6859a67c57fabbf4f"} Oct 09 10:24:55 crc kubenswrapper[4923]: I1009 10:24:55.312125 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 10:24:55 crc kubenswrapper[4923]: I1009 10:24:55.312616 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="c94c527a-e960-451f-87a8-4ed58af17475" containerName="kube-state-metrics" containerID="cri-o://e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738" gracePeriod=30 Oct 09 10:24:55 crc kubenswrapper[4923]: I1009 10:24:55.783179 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 10:24:55 crc kubenswrapper[4923]: I1009 10:24:55.909499 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxs2d\" (UniqueName: \"kubernetes.io/projected/c94c527a-e960-451f-87a8-4ed58af17475-kube-api-access-bxs2d\") pod \"c94c527a-e960-451f-87a8-4ed58af17475\" (UID: \"c94c527a-e960-451f-87a8-4ed58af17475\") " Oct 09 10:24:55 crc kubenswrapper[4923]: I1009 10:24:55.918062 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c94c527a-e960-451f-87a8-4ed58af17475-kube-api-access-bxs2d" (OuterVolumeSpecName: "kube-api-access-bxs2d") pod "c94c527a-e960-451f-87a8-4ed58af17475" (UID: "c94c527a-e960-451f-87a8-4ed58af17475"). InnerVolumeSpecName "kube-api-access-bxs2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.012887 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxs2d\" (UniqueName: \"kubernetes.io/projected/c94c527a-e960-451f-87a8-4ed58af17475-kube-api-access-bxs2d\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.230493 4923 generic.go:334] "Generic (PLEG): container finished" podID="c94c527a-e960-451f-87a8-4ed58af17475" containerID="e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738" exitCode=2 Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.230592 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.230626 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c94c527a-e960-451f-87a8-4ed58af17475","Type":"ContainerDied","Data":"e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738"} Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.231054 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c94c527a-e960-451f-87a8-4ed58af17475","Type":"ContainerDied","Data":"68cdc099aaf0cae2a0d152b4c401d7ad8ca27f0fe82f793ac3e2e11b741c1723"} Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.231102 4923 scope.go:117] "RemoveContainer" containerID="e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.233225 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"3f52a73f-7675-4cd5-ae86-77bd49333940","Type":"ContainerStarted","Data":"3917f5fdea39fb4ead104742136f0376696fabb8a5b97a508aee6ca6ebbf0618"} Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.233578 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.262447 4923 scope.go:117] "RemoveContainer" containerID="e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.264173 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.264148635 podStartE2EDuration="2.264148635s" podCreationTimestamp="2025-10-09 10:24:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:24:56.255435536 +0000 UTC m=+1182.323617292" watchObservedRunningTime="2025-10-09 10:24:56.264148635 +0000 UTC m=+1182.332330391" Oct 09 10:24:56 crc kubenswrapper[4923]: E1009 10:24:56.265356 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738\": container with ID starting with e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738 not found: ID does not exist" containerID="e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.265415 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738"} err="failed to get container status \"e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738\": rpc error: code = NotFound desc = could not find container \"e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738\": container with ID starting with e467cbbd22606678c7ef49ac4c2ae72fafbb0da2a32cf3cde982867fe00ad738 not found: ID does not exist" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.294657 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.306159 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.324934 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 10:24:56 crc kubenswrapper[4923]: E1009 10:24:56.325615 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c94c527a-e960-451f-87a8-4ed58af17475" containerName="kube-state-metrics" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.325641 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c94c527a-e960-451f-87a8-4ed58af17475" containerName="kube-state-metrics" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.325873 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c94c527a-e960-451f-87a8-4ed58af17475" containerName="kube-state-metrics" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.326665 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.333460 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.334656 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.349591 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.364804 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.365271 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="ceilometer-central-agent" containerID="cri-o://d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3" gracePeriod=30 Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.365837 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="proxy-httpd" containerID="cri-o://f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922" gracePeriod=30 Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.365973 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="ceilometer-notification-agent" containerID="cri-o://dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6" gracePeriod=30 Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.366058 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="sg-core" containerID="cri-o://3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372" gracePeriod=30 Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.422581 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d694c1-4a65-457d-8033-8b829df1ccbb-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.422883 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d694c1-4a65-457d-8033-8b829df1ccbb-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.423169 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmlfq\" (UniqueName: \"kubernetes.io/projected/73d694c1-4a65-457d-8033-8b829df1ccbb-kube-api-access-vmlfq\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.423295 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/73d694c1-4a65-457d-8033-8b829df1ccbb-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.526197 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d694c1-4a65-457d-8033-8b829df1ccbb-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.526361 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d694c1-4a65-457d-8033-8b829df1ccbb-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.526413 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmlfq\" (UniqueName: \"kubernetes.io/projected/73d694c1-4a65-457d-8033-8b829df1ccbb-kube-api-access-vmlfq\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.526447 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/73d694c1-4a65-457d-8033-8b829df1ccbb-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.533483 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/73d694c1-4a65-457d-8033-8b829df1ccbb-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.533783 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/73d694c1-4a65-457d-8033-8b829df1ccbb-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.538126 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73d694c1-4a65-457d-8033-8b829df1ccbb-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.552205 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmlfq\" (UniqueName: \"kubernetes.io/projected/73d694c1-4a65-457d-8033-8b829df1ccbb-kube-api-access-vmlfq\") pod \"kube-state-metrics-0\" (UID: \"73d694c1-4a65-457d-8033-8b829df1ccbb\") " pod="openstack/kube-state-metrics-0" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.612328 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c94c527a-e960-451f-87a8-4ed58af17475" path="/var/lib/kubelet/pods/c94c527a-e960-451f-87a8-4ed58af17475/volumes" Oct 09 10:24:56 crc kubenswrapper[4923]: I1009 10:24:56.653264 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 10:24:57 crc kubenswrapper[4923]: I1009 10:24:57.151482 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 10:24:57 crc kubenswrapper[4923]: I1009 10:24:57.246779 4923 generic.go:334] "Generic (PLEG): container finished" podID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerID="f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922" exitCode=0 Oct 09 10:24:57 crc kubenswrapper[4923]: I1009 10:24:57.246832 4923 generic.go:334] "Generic (PLEG): container finished" podID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerID="3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372" exitCode=2 Oct 09 10:24:57 crc kubenswrapper[4923]: I1009 10:24:57.246845 4923 generic.go:334] "Generic (PLEG): container finished" podID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerID="d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3" exitCode=0 Oct 09 10:24:57 crc kubenswrapper[4923]: I1009 10:24:57.246940 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerDied","Data":"f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922"} Oct 09 10:24:57 crc kubenswrapper[4923]: I1009 10:24:57.247210 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerDied","Data":"3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372"} Oct 09 10:24:57 crc kubenswrapper[4923]: I1009 10:24:57.247231 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerDied","Data":"d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3"} Oct 09 10:24:57 crc kubenswrapper[4923]: I1009 10:24:57.249022 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"73d694c1-4a65-457d-8033-8b829df1ccbb","Type":"ContainerStarted","Data":"e296181cc4b6f6ed52c3b8cb9758ccd387cdc2a992df7531ff19d1f8ea75be9f"} Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.026611 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.057457 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-scripts\") pod \"e6bd9476-77c4-45af-a38d-6972483c61c1\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.057607 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75gqp\" (UniqueName: \"kubernetes.io/projected/e6bd9476-77c4-45af-a38d-6972483c61c1-kube-api-access-75gqp\") pod \"e6bd9476-77c4-45af-a38d-6972483c61c1\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.057635 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-log-httpd\") pod \"e6bd9476-77c4-45af-a38d-6972483c61c1\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.057770 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-combined-ca-bundle\") pod \"e6bd9476-77c4-45af-a38d-6972483c61c1\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.057785 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-config-data\") pod \"e6bd9476-77c4-45af-a38d-6972483c61c1\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.057873 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-sg-core-conf-yaml\") pod \"e6bd9476-77c4-45af-a38d-6972483c61c1\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.057900 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-run-httpd\") pod \"e6bd9476-77c4-45af-a38d-6972483c61c1\" (UID: \"e6bd9476-77c4-45af-a38d-6972483c61c1\") " Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.058339 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e6bd9476-77c4-45af-a38d-6972483c61c1" (UID: "e6bd9476-77c4-45af-a38d-6972483c61c1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.058334 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e6bd9476-77c4-45af-a38d-6972483c61c1" (UID: "e6bd9476-77c4-45af-a38d-6972483c61c1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.065855 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6bd9476-77c4-45af-a38d-6972483c61c1-kube-api-access-75gqp" (OuterVolumeSpecName: "kube-api-access-75gqp") pod "e6bd9476-77c4-45af-a38d-6972483c61c1" (UID: "e6bd9476-77c4-45af-a38d-6972483c61c1"). InnerVolumeSpecName "kube-api-access-75gqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.067148 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-scripts" (OuterVolumeSpecName: "scripts") pod "e6bd9476-77c4-45af-a38d-6972483c61c1" (UID: "e6bd9476-77c4-45af-a38d-6972483c61c1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.116693 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e6bd9476-77c4-45af-a38d-6972483c61c1" (UID: "e6bd9476-77c4-45af-a38d-6972483c61c1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.139869 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e6bd9476-77c4-45af-a38d-6972483c61c1" (UID: "e6bd9476-77c4-45af-a38d-6972483c61c1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.160136 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.160167 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75gqp\" (UniqueName: \"kubernetes.io/projected/e6bd9476-77c4-45af-a38d-6972483c61c1-kube-api-access-75gqp\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.160181 4923 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.160191 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.160202 4923 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.160212 4923 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e6bd9476-77c4-45af-a38d-6972483c61c1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.177929 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-config-data" (OuterVolumeSpecName: "config-data") pod "e6bd9476-77c4-45af-a38d-6972483c61c1" (UID: "e6bd9476-77c4-45af-a38d-6972483c61c1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.262747 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e6bd9476-77c4-45af-a38d-6972483c61c1-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.265651 4923 generic.go:334] "Generic (PLEG): container finished" podID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerID="dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6" exitCode=0 Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.265739 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.265730 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerDied","Data":"dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6"} Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.265893 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e6bd9476-77c4-45af-a38d-6972483c61c1","Type":"ContainerDied","Data":"b2fa7110582e9356adc4c0b6657b9b437ed4aa3b101c8b141713b33b97577915"} Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.265939 4923 scope.go:117] "RemoveContainer" containerID="f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.268977 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"73d694c1-4a65-457d-8033-8b829df1ccbb","Type":"ContainerStarted","Data":"fbfadd487b468605e7ddc394f0bda778931ae9d4cc9f3d37c825b78756435ecf"} Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.269135 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.292830 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.7007270509999999 podStartE2EDuration="2.292809415s" podCreationTimestamp="2025-10-09 10:24:56 +0000 UTC" firstStartedPulling="2025-10-09 10:24:57.17316217 +0000 UTC m=+1183.241343926" lastFinishedPulling="2025-10-09 10:24:57.765244524 +0000 UTC m=+1183.833426290" observedRunningTime="2025-10-09 10:24:58.290182072 +0000 UTC m=+1184.358363848" watchObservedRunningTime="2025-10-09 10:24:58.292809415 +0000 UTC m=+1184.360991171" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.297616 4923 scope.go:117] "RemoveContainer" containerID="3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.320122 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.339497 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.346992 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:58 crc kubenswrapper[4923]: E1009 10:24:58.347547 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="ceilometer-notification-agent" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.347575 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="ceilometer-notification-agent" Oct 09 10:24:58 crc kubenswrapper[4923]: E1009 10:24:58.347613 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="sg-core" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.347622 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="sg-core" Oct 09 10:24:58 crc kubenswrapper[4923]: E1009 10:24:58.347641 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="proxy-httpd" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.347649 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="proxy-httpd" Oct 09 10:24:58 crc kubenswrapper[4923]: E1009 10:24:58.347675 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="ceilometer-central-agent" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.347682 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="ceilometer-central-agent" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.348002 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="proxy-httpd" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.348031 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="ceilometer-notification-agent" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.348059 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="sg-core" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.348074 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" containerName="ceilometer-central-agent" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.351264 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.353543 4923 scope.go:117] "RemoveContainer" containerID="dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.356181 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.356489 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.356673 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.366958 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.388713 4923 scope.go:117] "RemoveContainer" containerID="d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.412156 4923 scope.go:117] "RemoveContainer" containerID="f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922" Oct 09 10:24:58 crc kubenswrapper[4923]: E1009 10:24:58.413016 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922\": container with ID starting with f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922 not found: ID does not exist" containerID="f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.413074 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922"} err="failed to get container status \"f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922\": rpc error: code = NotFound desc = could not find container \"f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922\": container with ID starting with f504339b77d557840a9b3b65b2ee1e66e96c01f8c71f50f774855c6fd08e1922 not found: ID does not exist" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.413121 4923 scope.go:117] "RemoveContainer" containerID="3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372" Oct 09 10:24:58 crc kubenswrapper[4923]: E1009 10:24:58.413623 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372\": container with ID starting with 3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372 not found: ID does not exist" containerID="3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.413692 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372"} err="failed to get container status \"3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372\": rpc error: code = NotFound desc = could not find container \"3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372\": container with ID starting with 3e9c726090f2119d94064a39e8e5bdec410585a3531af1779f667d94abcb0372 not found: ID does not exist" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.413740 4923 scope.go:117] "RemoveContainer" containerID="dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6" Oct 09 10:24:58 crc kubenswrapper[4923]: E1009 10:24:58.414208 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6\": container with ID starting with dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6 not found: ID does not exist" containerID="dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.414280 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6"} err="failed to get container status \"dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6\": rpc error: code = NotFound desc = could not find container \"dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6\": container with ID starting with dbfaad61bcf2b47951a2123e7b3c8da20715eff6ae2075b557a7cab53a3646f6 not found: ID does not exist" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.414310 4923 scope.go:117] "RemoveContainer" containerID="d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3" Oct 09 10:24:58 crc kubenswrapper[4923]: E1009 10:24:58.415846 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3\": container with ID starting with d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3 not found: ID does not exist" containerID="d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.415897 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3"} err="failed to get container status \"d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3\": rpc error: code = NotFound desc = could not find container \"d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3\": container with ID starting with d6a010d212428219aa3d7f4018ab50db3d390d9534ca48a665cb2ca3b680bac3 not found: ID does not exist" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.466487 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-scripts\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.466722 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.466885 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.467015 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-run-httpd\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.467309 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.467467 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxzhv\" (UniqueName: \"kubernetes.io/projected/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-kube-api-access-zxzhv\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.467511 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-log-httpd\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.467672 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-config-data\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.570241 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-scripts\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.570367 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.570600 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.570649 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-run-httpd\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.570736 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.571158 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxzhv\" (UniqueName: \"kubernetes.io/projected/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-kube-api-access-zxzhv\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.571182 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-run-httpd\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.571197 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-log-httpd\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.571279 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-config-data\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.571497 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-log-httpd\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.574527 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.574862 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.575616 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.575936 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-config-data\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.576030 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-scripts\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.592482 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxzhv\" (UniqueName: \"kubernetes.io/projected/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-kube-api-access-zxzhv\") pod \"ceilometer-0\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " pod="openstack/ceilometer-0" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.614547 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6bd9476-77c4-45af-a38d-6972483c61c1" path="/var/lib/kubelet/pods/e6bd9476-77c4-45af-a38d-6972483c61c1/volumes" Oct 09 10:24:58 crc kubenswrapper[4923]: I1009 10:24:58.686190 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:24:59 crc kubenswrapper[4923]: I1009 10:24:59.145542 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:24:59 crc kubenswrapper[4923]: W1009 10:24:59.153155 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3214d04_5a88_4f5a_bb3a_9c3d172981ea.slice/crio-67119f01b6905ea3287bc7bddc5652c242bbc14ba90c4dd3ddd3ee630231339d WatchSource:0}: Error finding container 67119f01b6905ea3287bc7bddc5652c242bbc14ba90c4dd3ddd3ee630231339d: Status 404 returned error can't find the container with id 67119f01b6905ea3287bc7bddc5652c242bbc14ba90c4dd3ddd3ee630231339d Oct 09 10:24:59 crc kubenswrapper[4923]: I1009 10:24:59.281887 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerStarted","Data":"67119f01b6905ea3287bc7bddc5652c242bbc14ba90c4dd3ddd3ee630231339d"} Oct 09 10:25:01 crc kubenswrapper[4923]: I1009 10:25:01.330741 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerStarted","Data":"95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625"} Oct 09 10:25:02 crc kubenswrapper[4923]: I1009 10:25:02.340964 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerStarted","Data":"a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d"} Oct 09 10:25:02 crc kubenswrapper[4923]: I1009 10:25:02.341731 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerStarted","Data":"21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d"} Oct 09 10:25:04 crc kubenswrapper[4923]: I1009 10:25:04.364685 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerStarted","Data":"0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859"} Oct 09 10:25:04 crc kubenswrapper[4923]: I1009 10:25:04.367629 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 10:25:04 crc kubenswrapper[4923]: I1009 10:25:04.394887 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.217127946 podStartE2EDuration="6.394862692s" podCreationTimestamp="2025-10-09 10:24:58 +0000 UTC" firstStartedPulling="2025-10-09 10:24:59.156657718 +0000 UTC m=+1185.224839484" lastFinishedPulling="2025-10-09 10:25:03.334392474 +0000 UTC m=+1189.402574230" observedRunningTime="2025-10-09 10:25:04.389074023 +0000 UTC m=+1190.457255779" watchObservedRunningTime="2025-10-09 10:25:04.394862692 +0000 UTC m=+1190.463044448" Oct 09 10:25:04 crc kubenswrapper[4923]: I1009 10:25:04.677292 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.165819 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-hkxgk"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.168265 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.170975 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.171373 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.177034 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hkxgk"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.226849 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-config-data\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.227047 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-scripts\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.227081 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.227122 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkx9s\" (UniqueName: \"kubernetes.io/projected/5c4e9807-4e89-49d0-bea0-732464e07ca2-kube-api-access-xkx9s\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.340152 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-config-data\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.340305 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-scripts\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.340333 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.340351 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkx9s\" (UniqueName: \"kubernetes.io/projected/5c4e9807-4e89-49d0-bea0-732464e07ca2-kube-api-access-xkx9s\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.351809 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.383426 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-config-data\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.400243 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-scripts\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.403997 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkx9s\" (UniqueName: \"kubernetes.io/projected/5c4e9807-4e89-49d0-bea0-732464e07ca2-kube-api-access-xkx9s\") pod \"nova-cell0-cell-mapping-hkxgk\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.405614 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.424679 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.445676 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.464426 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.482513 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.485152 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.503893 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.504381 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.541844 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.556096 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.556165 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-config-data\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.556272 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.556294 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm5fl\" (UniqueName: \"kubernetes.io/projected/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-kube-api-access-xm5fl\") pod \"nova-scheduler-0\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.556387 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-logs\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.556430 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-config-data\") pod \"nova-scheduler-0\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.556490 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czbwz\" (UniqueName: \"kubernetes.io/projected/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-kube-api-access-czbwz\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.577847 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.579520 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.604264 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.633554 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665203 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czbwz\" (UniqueName: \"kubernetes.io/projected/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-kube-api-access-czbwz\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665277 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rddp\" (UniqueName: \"kubernetes.io/projected/ed966e00-bc66-4579-b9b1-e16c7dd82854-kube-api-access-9rddp\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665398 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665428 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-config-data\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665466 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665550 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665576 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm5fl\" (UniqueName: \"kubernetes.io/projected/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-kube-api-access-xm5fl\") pod \"nova-scheduler-0\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665713 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-logs\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665802 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-config-data\") pod \"nova-scheduler-0\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.665931 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.668944 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-logs\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.683211 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-config-data\") pod \"nova-scheduler-0\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.684328 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.703795 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-config-data\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.728728 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm5fl\" (UniqueName: \"kubernetes.io/projected/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-kube-api-access-xm5fl\") pod \"nova-scheduler-0\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.742369 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czbwz\" (UniqueName: \"kubernetes.io/projected/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-kube-api-access-czbwz\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.742908 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.769295 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.769499 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.769535 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rddp\" (UniqueName: \"kubernetes.io/projected/ed966e00-bc66-4579-b9b1-e16c7dd82854-kube-api-access-9rddp\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.788401 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.789130 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.797403 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.800588 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.812545 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.820049 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.846809 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rddp\" (UniqueName: \"kubernetes.io/projected/ed966e00-bc66-4579-b9b1-e16c7dd82854-kube-api-access-9rddp\") pod \"nova-cell1-novncproxy-0\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.856866 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.887445 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.982181 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-config-data\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.982285 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.982357 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/682656b7-314d-4568-a105-630ff23b6678-logs\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.982480 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqhtx\" (UniqueName: \"kubernetes.io/projected/682656b7-314d-4568-a105-630ff23b6678-kube-api-access-sqhtx\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.983577 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-9scnz"] Oct 09 10:25:05 crc kubenswrapper[4923]: I1009 10:25:05.997792 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.032284 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-9scnz"] Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.081875 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.083597 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqhtx\" (UniqueName: \"kubernetes.io/projected/682656b7-314d-4568-a105-630ff23b6678-kube-api-access-sqhtx\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.083641 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-config-data\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.083683 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.083741 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/682656b7-314d-4568-a105-630ff23b6678-logs\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.086597 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/682656b7-314d-4568-a105-630ff23b6678-logs\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.094563 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.108556 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-config-data\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.116400 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqhtx\" (UniqueName: \"kubernetes.io/projected/682656b7-314d-4568-a105-630ff23b6678-kube-api-access-sqhtx\") pod \"nova-metadata-0\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " pod="openstack/nova-metadata-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.190534 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.191229 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpbdd\" (UniqueName: \"kubernetes.io/projected/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-kube-api-access-gpbdd\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.191299 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-config\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.191348 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-nb\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.191672 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-sb\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.191847 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-dns-svc\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.295665 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpbdd\" (UniqueName: \"kubernetes.io/projected/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-kube-api-access-gpbdd\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.296417 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-config\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.296601 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-nb\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.296694 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-sb\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.296825 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-dns-svc\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.298431 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-nb\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.298998 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-config\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.299472 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-sb\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.300435 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-dns-svc\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.332291 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpbdd\" (UniqueName: \"kubernetes.io/projected/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-kube-api-access-gpbdd\") pod \"dnsmasq-dns-54974c8ff5-9scnz\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.364499 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.441512 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hkxgk"] Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.588563 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.598530 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.700394 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.783045 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-78qxk"] Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.784600 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.790104 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.798299 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.817115 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-78qxk"] Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.896919 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.917434 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-scripts\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.917490 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.917595 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs4hs\" (UniqueName: \"kubernetes.io/projected/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-kube-api-access-rs4hs\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.917783 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-config-data\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:06 crc kubenswrapper[4923]: I1009 10:25:06.935596 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.022277 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-scripts\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.022352 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.022373 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs4hs\" (UniqueName: \"kubernetes.io/projected/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-kube-api-access-rs4hs\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.022623 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-config-data\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.030262 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-scripts\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.032208 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.033017 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-config-data\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.042198 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-9scnz"] Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.049319 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs4hs\" (UniqueName: \"kubernetes.io/projected/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-kube-api-access-rs4hs\") pod \"nova-cell1-conductor-db-sync-78qxk\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.145553 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.481705 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hkxgk" event={"ID":"5c4e9807-4e89-49d0-bea0-732464e07ca2","Type":"ContainerStarted","Data":"05d08cf8c95399fd1938b6d73b8d94427366df1682385eba5b83ce0c1b34e101"} Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.482100 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hkxgk" event={"ID":"5c4e9807-4e89-49d0-bea0-732464e07ca2","Type":"ContainerStarted","Data":"29bb85ae9c1483d7f4bf928aa8a9c2cf2f63874208e38c931587151ccb233ba2"} Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.485501 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"682656b7-314d-4568-a105-630ff23b6678","Type":"ContainerStarted","Data":"e99343a1c77bffb367cd35a783eb90ab486dcdb3d610d6a3006c5777b4903757"} Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.493971 4923 generic.go:334] "Generic (PLEG): container finished" podID="6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" containerID="40fc71db54b599f418267f1ad1da8f30e84e7b8de3bc8701928ee0228a83ba54" exitCode=0 Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.494074 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" event={"ID":"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6","Type":"ContainerDied","Data":"40fc71db54b599f418267f1ad1da8f30e84e7b8de3bc8701928ee0228a83ba54"} Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.494172 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" event={"ID":"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6","Type":"ContainerStarted","Data":"c2f2f326f7121fbbab6af0b1dc1c82bc8697570485e54b944a8b2d1252a7e62b"} Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.497031 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed966e00-bc66-4579-b9b1-e16c7dd82854","Type":"ContainerStarted","Data":"997c76ee6a3cb56c92e3a94cd2e3702a72b6ed923a52ff05ce2bc239ba17d374"} Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.500104 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a111f6-7635-4a6d-9aab-fda23fcf4c54","Type":"ContainerStarted","Data":"fe0c69f10457181d9a277223450b807c8568248e66d81cd7e84b149d093f0bb9"} Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.506825 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-hkxgk" podStartSLOduration=2.506791125 podStartE2EDuration="2.506791125s" podCreationTimestamp="2025-10-09 10:25:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:07.500576114 +0000 UTC m=+1193.568757880" watchObservedRunningTime="2025-10-09 10:25:07.506791125 +0000 UTC m=+1193.574972901" Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.511826 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"473dfb33-5dd7-4805-a57d-73fffe7e9bcd","Type":"ContainerStarted","Data":"01c50b747d8369ed16534712ed820ecde0394e32706e6569dba9307feeaf1a32"} Oct 09 10:25:07 crc kubenswrapper[4923]: I1009 10:25:07.656710 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-78qxk"] Oct 09 10:25:08 crc kubenswrapper[4923]: I1009 10:25:08.531479 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-78qxk" event={"ID":"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1","Type":"ContainerStarted","Data":"9d7fbf539527d18ed86c2341708e436a44ee2c29f3a95f9f99732acbacd9422b"} Oct 09 10:25:08 crc kubenswrapper[4923]: I1009 10:25:08.532429 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-78qxk" event={"ID":"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1","Type":"ContainerStarted","Data":"50bad4991bc744981af2567c75fc935ada0370130df332dc5c36a8b4b0f35973"} Oct 09 10:25:08 crc kubenswrapper[4923]: I1009 10:25:08.539870 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" event={"ID":"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6","Type":"ContainerStarted","Data":"19ced66d3a8413134b8e05e7b359f29a4b16c6b7ce0d1861cb75903f98fe986b"} Oct 09 10:25:08 crc kubenswrapper[4923]: I1009 10:25:08.541057 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:08 crc kubenswrapper[4923]: I1009 10:25:08.578701 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-78qxk" podStartSLOduration=2.578666916 podStartE2EDuration="2.578666916s" podCreationTimestamp="2025-10-09 10:25:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:08.577132664 +0000 UTC m=+1194.645314650" watchObservedRunningTime="2025-10-09 10:25:08.578666916 +0000 UTC m=+1194.646848662" Oct 09 10:25:08 crc kubenswrapper[4923]: I1009 10:25:08.611349 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" podStartSLOduration=3.611320014 podStartE2EDuration="3.611320014s" podCreationTimestamp="2025-10-09 10:25:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:08.603613401 +0000 UTC m=+1194.671795177" watchObservedRunningTime="2025-10-09 10:25:08.611320014 +0000 UTC m=+1194.679501770" Oct 09 10:25:09 crc kubenswrapper[4923]: I1009 10:25:09.377638 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 10:25:09 crc kubenswrapper[4923]: I1009 10:25:09.421694 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.595542 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a111f6-7635-4a6d-9aab-fda23fcf4c54","Type":"ContainerStarted","Data":"837f0b14e2ef8b5458a707e89f723b9ff26d292f529528208fe839a167eab602"} Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.596474 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a111f6-7635-4a6d-9aab-fda23fcf4c54","Type":"ContainerStarted","Data":"4ef03f06e925dbea844c1fb919b1c69d81d41ea701cce5de9b30356816c08f59"} Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.599473 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"473dfb33-5dd7-4805-a57d-73fffe7e9bcd","Type":"ContainerStarted","Data":"4a90b2287c8ad383ed78f2913ee13f17eb04c81228bfb72a0e41822f619e498c"} Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.602284 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="682656b7-314d-4568-a105-630ff23b6678" containerName="nova-metadata-log" containerID="cri-o://95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1" gracePeriod=30 Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.602435 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="682656b7-314d-4568-a105-630ff23b6678" containerName="nova-metadata-metadata" containerID="cri-o://d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d" gracePeriod=30 Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.604764 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ed966e00-bc66-4579-b9b1-e16c7dd82854" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d" gracePeriod=30 Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.620904 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"682656b7-314d-4568-a105-630ff23b6678","Type":"ContainerStarted","Data":"d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d"} Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.620969 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"682656b7-314d-4568-a105-630ff23b6678","Type":"ContainerStarted","Data":"95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1"} Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.620986 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed966e00-bc66-4579-b9b1-e16c7dd82854","Type":"ContainerStarted","Data":"80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d"} Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.629921 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.037428838 podStartE2EDuration="7.629886765s" podCreationTimestamp="2025-10-09 10:25:05 +0000 UTC" firstStartedPulling="2025-10-09 10:25:06.617646656 +0000 UTC m=+1192.685828412" lastFinishedPulling="2025-10-09 10:25:11.210104583 +0000 UTC m=+1197.278286339" observedRunningTime="2025-10-09 10:25:12.6155012 +0000 UTC m=+1198.683682966" watchObservedRunningTime="2025-10-09 10:25:12.629886765 +0000 UTC m=+1198.698068521" Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.640096 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.365836417 podStartE2EDuration="7.640070466s" podCreationTimestamp="2025-10-09 10:25:05 +0000 UTC" firstStartedPulling="2025-10-09 10:25:06.946685141 +0000 UTC m=+1193.014866897" lastFinishedPulling="2025-10-09 10:25:11.22091919 +0000 UTC m=+1197.289100946" observedRunningTime="2025-10-09 10:25:12.634722219 +0000 UTC m=+1198.702903995" watchObservedRunningTime="2025-10-09 10:25:12.640070466 +0000 UTC m=+1198.708252222" Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.664920 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.321702193 podStartE2EDuration="7.664894318s" podCreationTimestamp="2025-10-09 10:25:05 +0000 UTC" firstStartedPulling="2025-10-09 10:25:06.867692279 +0000 UTC m=+1192.935874035" lastFinishedPulling="2025-10-09 10:25:11.210884414 +0000 UTC m=+1197.279066160" observedRunningTime="2025-10-09 10:25:12.659315735 +0000 UTC m=+1198.727497491" watchObservedRunningTime="2025-10-09 10:25:12.664894318 +0000 UTC m=+1198.733076074" Oct 09 10:25:12 crc kubenswrapper[4923]: I1009 10:25:12.687098 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.122008135 podStartE2EDuration="7.687066468s" podCreationTimestamp="2025-10-09 10:25:05 +0000 UTC" firstStartedPulling="2025-10-09 10:25:06.644917796 +0000 UTC m=+1192.713099552" lastFinishedPulling="2025-10-09 10:25:11.209976129 +0000 UTC m=+1197.278157885" observedRunningTime="2025-10-09 10:25:12.68317513 +0000 UTC m=+1198.751356906" watchObservedRunningTime="2025-10-09 10:25:12.687066468 +0000 UTC m=+1198.755248244" Oct 09 10:25:13 crc kubenswrapper[4923]: I1009 10:25:13.616892 4923 generic.go:334] "Generic (PLEG): container finished" podID="682656b7-314d-4568-a105-630ff23b6678" containerID="95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1" exitCode=143 Oct 09 10:25:13 crc kubenswrapper[4923]: I1009 10:25:13.616976 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"682656b7-314d-4568-a105-630ff23b6678","Type":"ContainerDied","Data":"95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1"} Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.614856 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.642872 4923 generic.go:334] "Generic (PLEG): container finished" podID="682656b7-314d-4568-a105-630ff23b6678" containerID="d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d" exitCode=0 Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.642945 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"682656b7-314d-4568-a105-630ff23b6678","Type":"ContainerDied","Data":"d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d"} Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.642995 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"682656b7-314d-4568-a105-630ff23b6678","Type":"ContainerDied","Data":"e99343a1c77bffb367cd35a783eb90ab486dcdb3d610d6a3006c5777b4903757"} Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.643042 4923 scope.go:117] "RemoveContainer" containerID="d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.643264 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.681574 4923 scope.go:117] "RemoveContainer" containerID="95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.710146 4923 scope.go:117] "RemoveContainer" containerID="d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d" Oct 09 10:25:14 crc kubenswrapper[4923]: E1009 10:25:14.710768 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d\": container with ID starting with d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d not found: ID does not exist" containerID="d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.710854 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d"} err="failed to get container status \"d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d\": rpc error: code = NotFound desc = could not find container \"d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d\": container with ID starting with d79776f032d4aa67a49274427fb7b07ac60112287882ddb9bd4bebcfc94b0f6d not found: ID does not exist" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.710909 4923 scope.go:117] "RemoveContainer" containerID="95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1" Oct 09 10:25:14 crc kubenswrapper[4923]: E1009 10:25:14.711945 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1\": container with ID starting with 95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1 not found: ID does not exist" containerID="95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.712024 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1"} err="failed to get container status \"95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1\": rpc error: code = NotFound desc = could not find container \"95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1\": container with ID starting with 95881dcbc41f608fcbc8b21e5237d7e415f491502d93644a368730b69e1859a1 not found: ID does not exist" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.739514 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-config-data\") pod \"682656b7-314d-4568-a105-630ff23b6678\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.739614 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/682656b7-314d-4568-a105-630ff23b6678-logs\") pod \"682656b7-314d-4568-a105-630ff23b6678\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.739775 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqhtx\" (UniqueName: \"kubernetes.io/projected/682656b7-314d-4568-a105-630ff23b6678-kube-api-access-sqhtx\") pod \"682656b7-314d-4568-a105-630ff23b6678\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.739966 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-combined-ca-bundle\") pod \"682656b7-314d-4568-a105-630ff23b6678\" (UID: \"682656b7-314d-4568-a105-630ff23b6678\") " Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.740321 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/682656b7-314d-4568-a105-630ff23b6678-logs" (OuterVolumeSpecName: "logs") pod "682656b7-314d-4568-a105-630ff23b6678" (UID: "682656b7-314d-4568-a105-630ff23b6678"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.740716 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/682656b7-314d-4568-a105-630ff23b6678-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.753856 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/682656b7-314d-4568-a105-630ff23b6678-kube-api-access-sqhtx" (OuterVolumeSpecName: "kube-api-access-sqhtx") pod "682656b7-314d-4568-a105-630ff23b6678" (UID: "682656b7-314d-4568-a105-630ff23b6678"). InnerVolumeSpecName "kube-api-access-sqhtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.776394 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-config-data" (OuterVolumeSpecName: "config-data") pod "682656b7-314d-4568-a105-630ff23b6678" (UID: "682656b7-314d-4568-a105-630ff23b6678"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.783819 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "682656b7-314d-4568-a105-630ff23b6678" (UID: "682656b7-314d-4568-a105-630ff23b6678"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.842722 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.842794 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/682656b7-314d-4568-a105-630ff23b6678-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.842809 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqhtx\" (UniqueName: \"kubernetes.io/projected/682656b7-314d-4568-a105-630ff23b6678-kube-api-access-sqhtx\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.980891 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:14 crc kubenswrapper[4923]: I1009 10:25:14.991226 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.013571 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:15 crc kubenswrapper[4923]: E1009 10:25:15.014385 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="682656b7-314d-4568-a105-630ff23b6678" containerName="nova-metadata-metadata" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.014416 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="682656b7-314d-4568-a105-630ff23b6678" containerName="nova-metadata-metadata" Oct 09 10:25:15 crc kubenswrapper[4923]: E1009 10:25:15.014442 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="682656b7-314d-4568-a105-630ff23b6678" containerName="nova-metadata-log" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.014451 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="682656b7-314d-4568-a105-630ff23b6678" containerName="nova-metadata-log" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.014660 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="682656b7-314d-4568-a105-630ff23b6678" containerName="nova-metadata-log" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.014690 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="682656b7-314d-4568-a105-630ff23b6678" containerName="nova-metadata-metadata" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.016007 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.019357 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.019361 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.022606 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.046567 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-config-data\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.046731 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/340330a2-dbec-4301-a703-7ab727aad3d9-logs\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.046784 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj4rv\" (UniqueName: \"kubernetes.io/projected/340330a2-dbec-4301-a703-7ab727aad3d9-kube-api-access-jj4rv\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.046821 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.046872 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.148413 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/340330a2-dbec-4301-a703-7ab727aad3d9-logs\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.148466 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj4rv\" (UniqueName: \"kubernetes.io/projected/340330a2-dbec-4301-a703-7ab727aad3d9-kube-api-access-jj4rv\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.148502 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.148544 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.148585 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-config-data\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.153373 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-config-data\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.153695 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/340330a2-dbec-4301-a703-7ab727aad3d9-logs\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.156464 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.157659 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.176433 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj4rv\" (UniqueName: \"kubernetes.io/projected/340330a2-dbec-4301-a703-7ab727aad3d9-kube-api-access-jj4rv\") pod \"nova-metadata-0\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.354967 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.822979 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.823425 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.838767 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.861897 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.862210 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 09 10:25:15 crc kubenswrapper[4923]: I1009 10:25:15.919019 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.083135 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.369591 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.445462 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-hlckn"] Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.445844 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85494b87f-hlckn" podUID="51933409-9af8-4cd7-8642-6d4f675068d7" containerName="dnsmasq-dns" containerID="cri-o://6a8a0d5cd2b6feda4d71058d60caadf91433abbd2a4b7b65a9cd7f1f63412af1" gracePeriod=10 Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.620345 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="682656b7-314d-4568-a105-630ff23b6678" path="/var/lib/kubelet/pods/682656b7-314d-4568-a105-630ff23b6678/volumes" Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.681622 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"340330a2-dbec-4301-a703-7ab727aad3d9","Type":"ContainerStarted","Data":"bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5"} Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.681690 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"340330a2-dbec-4301-a703-7ab727aad3d9","Type":"ContainerStarted","Data":"ae5bc6a3d9ee674c7d9360084f1c7cb4dc1e8a304f262a904f11205e217f0600"} Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.686814 4923 generic.go:334] "Generic (PLEG): container finished" podID="51933409-9af8-4cd7-8642-6d4f675068d7" containerID="6a8a0d5cd2b6feda4d71058d60caadf91433abbd2a4b7b65a9cd7f1f63412af1" exitCode=0 Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.686901 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-hlckn" event={"ID":"51933409-9af8-4cd7-8642-6d4f675068d7","Type":"ContainerDied","Data":"6a8a0d5cd2b6feda4d71058d60caadf91433abbd2a4b7b65a9cd7f1f63412af1"} Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.742059 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.907053 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.176:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 10:25:16 crc kubenswrapper[4923]: I1009 10:25:16.907702 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.176:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.159046 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.201887 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-sb\") pod \"51933409-9af8-4cd7-8642-6d4f675068d7\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.202317 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-config\") pod \"51933409-9af8-4cd7-8642-6d4f675068d7\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.202444 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2wl5\" (UniqueName: \"kubernetes.io/projected/51933409-9af8-4cd7-8642-6d4f675068d7-kube-api-access-b2wl5\") pod \"51933409-9af8-4cd7-8642-6d4f675068d7\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.202715 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-dns-svc\") pod \"51933409-9af8-4cd7-8642-6d4f675068d7\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.202829 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-nb\") pod \"51933409-9af8-4cd7-8642-6d4f675068d7\" (UID: \"51933409-9af8-4cd7-8642-6d4f675068d7\") " Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.280065 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/51933409-9af8-4cd7-8642-6d4f675068d7-kube-api-access-b2wl5" (OuterVolumeSpecName: "kube-api-access-b2wl5") pod "51933409-9af8-4cd7-8642-6d4f675068d7" (UID: "51933409-9af8-4cd7-8642-6d4f675068d7"). InnerVolumeSpecName "kube-api-access-b2wl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.311642 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2wl5\" (UniqueName: \"kubernetes.io/projected/51933409-9af8-4cd7-8642-6d4f675068d7-kube-api-access-b2wl5\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.342940 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "51933409-9af8-4cd7-8642-6d4f675068d7" (UID: "51933409-9af8-4cd7-8642-6d4f675068d7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.346610 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-config" (OuterVolumeSpecName: "config") pod "51933409-9af8-4cd7-8642-6d4f675068d7" (UID: "51933409-9af8-4cd7-8642-6d4f675068d7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.414524 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.415027 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.422523 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "51933409-9af8-4cd7-8642-6d4f675068d7" (UID: "51933409-9af8-4cd7-8642-6d4f675068d7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.423151 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "51933409-9af8-4cd7-8642-6d4f675068d7" (UID: "51933409-9af8-4cd7-8642-6d4f675068d7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.520411 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.520477 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/51933409-9af8-4cd7-8642-6d4f675068d7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.700277 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85494b87f-hlckn" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.700736 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85494b87f-hlckn" event={"ID":"51933409-9af8-4cd7-8642-6d4f675068d7","Type":"ContainerDied","Data":"136ea258b6bb1ff3fcd8d4d7251d283af686e9f96538eb629a4ab119af745cb9"} Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.700897 4923 scope.go:117] "RemoveContainer" containerID="6a8a0d5cd2b6feda4d71058d60caadf91433abbd2a4b7b65a9cd7f1f63412af1" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.709394 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"340330a2-dbec-4301-a703-7ab727aad3d9","Type":"ContainerStarted","Data":"59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63"} Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.730224 4923 scope.go:117] "RemoveContainer" containerID="df3599959384db95c3eef973f1475b7365cfd91649b14fc382f41be572fa9e83" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.762826 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.7627957260000002 podStartE2EDuration="3.762795726s" podCreationTimestamp="2025-10-09 10:25:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:17.731157286 +0000 UTC m=+1203.799339042" watchObservedRunningTime="2025-10-09 10:25:17.762795726 +0000 UTC m=+1203.830977482" Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.803514 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-hlckn"] Oct 09 10:25:17 crc kubenswrapper[4923]: I1009 10:25:17.822578 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85494b87f-hlckn"] Oct 09 10:25:18 crc kubenswrapper[4923]: I1009 10:25:18.614697 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="51933409-9af8-4cd7-8642-6d4f675068d7" path="/var/lib/kubelet/pods/51933409-9af8-4cd7-8642-6d4f675068d7/volumes" Oct 09 10:25:18 crc kubenswrapper[4923]: I1009 10:25:18.721861 4923 generic.go:334] "Generic (PLEG): container finished" podID="5c4e9807-4e89-49d0-bea0-732464e07ca2" containerID="05d08cf8c95399fd1938b6d73b8d94427366df1682385eba5b83ce0c1b34e101" exitCode=0 Oct 09 10:25:18 crc kubenswrapper[4923]: I1009 10:25:18.721947 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hkxgk" event={"ID":"5c4e9807-4e89-49d0-bea0-732464e07ca2","Type":"ContainerDied","Data":"05d08cf8c95399fd1938b6d73b8d94427366df1682385eba5b83ce0c1b34e101"} Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.130097 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.195801 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-scripts\") pod \"5c4e9807-4e89-49d0-bea0-732464e07ca2\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.197676 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkx9s\" (UniqueName: \"kubernetes.io/projected/5c4e9807-4e89-49d0-bea0-732464e07ca2-kube-api-access-xkx9s\") pod \"5c4e9807-4e89-49d0-bea0-732464e07ca2\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.197889 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-combined-ca-bundle\") pod \"5c4e9807-4e89-49d0-bea0-732464e07ca2\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.197937 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-config-data\") pod \"5c4e9807-4e89-49d0-bea0-732464e07ca2\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.204560 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c4e9807-4e89-49d0-bea0-732464e07ca2-kube-api-access-xkx9s" (OuterVolumeSpecName: "kube-api-access-xkx9s") pod "5c4e9807-4e89-49d0-bea0-732464e07ca2" (UID: "5c4e9807-4e89-49d0-bea0-732464e07ca2"). InnerVolumeSpecName "kube-api-access-xkx9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.204578 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-scripts" (OuterVolumeSpecName: "scripts") pod "5c4e9807-4e89-49d0-bea0-732464e07ca2" (UID: "5c4e9807-4e89-49d0-bea0-732464e07ca2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:20 crc kubenswrapper[4923]: E1009 10:25:20.230808 4923 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-config-data podName:5c4e9807-4e89-49d0-bea0-732464e07ca2 nodeName:}" failed. No retries permitted until 2025-10-09 10:25:20.730740539 +0000 UTC m=+1206.798922295 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-config-data") pod "5c4e9807-4e89-49d0-bea0-732464e07ca2" (UID: "5c4e9807-4e89-49d0-bea0-732464e07ca2") : error deleting /var/lib/kubelet/pods/5c4e9807-4e89-49d0-bea0-732464e07ca2/volume-subpaths: remove /var/lib/kubelet/pods/5c4e9807-4e89-49d0-bea0-732464e07ca2/volume-subpaths: no such file or directory Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.234252 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c4e9807-4e89-49d0-bea0-732464e07ca2" (UID: "5c4e9807-4e89-49d0-bea0-732464e07ca2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.300469 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkx9s\" (UniqueName: \"kubernetes.io/projected/5c4e9807-4e89-49d0-bea0-732464e07ca2-kube-api-access-xkx9s\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.300835 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.300930 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.355909 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.356312 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.744301 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hkxgk" event={"ID":"5c4e9807-4e89-49d0-bea0-732464e07ca2","Type":"ContainerDied","Data":"29bb85ae9c1483d7f4bf928aa8a9c2cf2f63874208e38c931587151ccb233ba2"} Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.744874 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29bb85ae9c1483d7f4bf928aa8a9c2cf2f63874208e38c931587151ccb233ba2" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.744359 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hkxgk" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.811367 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-config-data\") pod \"5c4e9807-4e89-49d0-bea0-732464e07ca2\" (UID: \"5c4e9807-4e89-49d0-bea0-732464e07ca2\") " Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.816461 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-config-data" (OuterVolumeSpecName: "config-data") pod "5c4e9807-4e89-49d0-bea0-732464e07ca2" (UID: "5c4e9807-4e89-49d0-bea0-732464e07ca2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.914709 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c4e9807-4e89-49d0-bea0-732464e07ca2-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.994710 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.995091 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-log" containerID="cri-o://4ef03f06e925dbea844c1fb919b1c69d81d41ea701cce5de9b30356816c08f59" gracePeriod=30 Oct 09 10:25:20 crc kubenswrapper[4923]: I1009 10:25:20.995202 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-api" containerID="cri-o://837f0b14e2ef8b5458a707e89f723b9ff26d292f529528208fe839a167eab602" gracePeriod=30 Oct 09 10:25:21 crc kubenswrapper[4923]: I1009 10:25:21.070608 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:21 crc kubenswrapper[4923]: I1009 10:25:21.115926 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:25:21 crc kubenswrapper[4923]: I1009 10:25:21.116193 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="473dfb33-5dd7-4805-a57d-73fffe7e9bcd" containerName="nova-scheduler-scheduler" containerID="cri-o://4a90b2287c8ad383ed78f2913ee13f17eb04c81228bfb72a0e41822f619e498c" gracePeriod=30 Oct 09 10:25:21 crc kubenswrapper[4923]: I1009 10:25:21.756063 4923 generic.go:334] "Generic (PLEG): container finished" podID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerID="4ef03f06e925dbea844c1fb919b1c69d81d41ea701cce5de9b30356816c08f59" exitCode=143 Oct 09 10:25:21 crc kubenswrapper[4923]: I1009 10:25:21.756159 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a111f6-7635-4a6d-9aab-fda23fcf4c54","Type":"ContainerDied","Data":"4ef03f06e925dbea844c1fb919b1c69d81d41ea701cce5de9b30356816c08f59"} Oct 09 10:25:21 crc kubenswrapper[4923]: I1009 10:25:21.756303 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="340330a2-dbec-4301-a703-7ab727aad3d9" containerName="nova-metadata-log" containerID="cri-o://bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5" gracePeriod=30 Oct 09 10:25:21 crc kubenswrapper[4923]: I1009 10:25:21.756840 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="340330a2-dbec-4301-a703-7ab727aad3d9" containerName="nova-metadata-metadata" containerID="cri-o://59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63" gracePeriod=30 Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.374605 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.446571 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/340330a2-dbec-4301-a703-7ab727aad3d9-logs\") pod \"340330a2-dbec-4301-a703-7ab727aad3d9\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.446732 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-config-data\") pod \"340330a2-dbec-4301-a703-7ab727aad3d9\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.447113 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-nova-metadata-tls-certs\") pod \"340330a2-dbec-4301-a703-7ab727aad3d9\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.447137 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-combined-ca-bundle\") pod \"340330a2-dbec-4301-a703-7ab727aad3d9\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.447259 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jj4rv\" (UniqueName: \"kubernetes.io/projected/340330a2-dbec-4301-a703-7ab727aad3d9-kube-api-access-jj4rv\") pod \"340330a2-dbec-4301-a703-7ab727aad3d9\" (UID: \"340330a2-dbec-4301-a703-7ab727aad3d9\") " Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.449031 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/340330a2-dbec-4301-a703-7ab727aad3d9-logs" (OuterVolumeSpecName: "logs") pod "340330a2-dbec-4301-a703-7ab727aad3d9" (UID: "340330a2-dbec-4301-a703-7ab727aad3d9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.456790 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/340330a2-dbec-4301-a703-7ab727aad3d9-kube-api-access-jj4rv" (OuterVolumeSpecName: "kube-api-access-jj4rv") pod "340330a2-dbec-4301-a703-7ab727aad3d9" (UID: "340330a2-dbec-4301-a703-7ab727aad3d9"). InnerVolumeSpecName "kube-api-access-jj4rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.487555 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-config-data" (OuterVolumeSpecName: "config-data") pod "340330a2-dbec-4301-a703-7ab727aad3d9" (UID: "340330a2-dbec-4301-a703-7ab727aad3d9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.493493 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "340330a2-dbec-4301-a703-7ab727aad3d9" (UID: "340330a2-dbec-4301-a703-7ab727aad3d9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.516993 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "340330a2-dbec-4301-a703-7ab727aad3d9" (UID: "340330a2-dbec-4301-a703-7ab727aad3d9"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.549639 4923 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.549678 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.549689 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jj4rv\" (UniqueName: \"kubernetes.io/projected/340330a2-dbec-4301-a703-7ab727aad3d9-kube-api-access-jj4rv\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.549711 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/340330a2-dbec-4301-a703-7ab727aad3d9-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.549722 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/340330a2-dbec-4301-a703-7ab727aad3d9-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.768259 4923 generic.go:334] "Generic (PLEG): container finished" podID="340330a2-dbec-4301-a703-7ab727aad3d9" containerID="59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63" exitCode=0 Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.768619 4923 generic.go:334] "Generic (PLEG): container finished" podID="340330a2-dbec-4301-a703-7ab727aad3d9" containerID="bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5" exitCode=143 Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.768362 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"340330a2-dbec-4301-a703-7ab727aad3d9","Type":"ContainerDied","Data":"59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63"} Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.768338 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.768691 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"340330a2-dbec-4301-a703-7ab727aad3d9","Type":"ContainerDied","Data":"bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5"} Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.768711 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"340330a2-dbec-4301-a703-7ab727aad3d9","Type":"ContainerDied","Data":"ae5bc6a3d9ee674c7d9360084f1c7cb4dc1e8a304f262a904f11205e217f0600"} Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.768737 4923 scope.go:117] "RemoveContainer" containerID="59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.796336 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.801048 4923 scope.go:117] "RemoveContainer" containerID="bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.810000 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827124 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:22 crc kubenswrapper[4923]: E1009 10:25:22.827572 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51933409-9af8-4cd7-8642-6d4f675068d7" containerName="init" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827591 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="51933409-9af8-4cd7-8642-6d4f675068d7" containerName="init" Oct 09 10:25:22 crc kubenswrapper[4923]: E1009 10:25:22.827600 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c4e9807-4e89-49d0-bea0-732464e07ca2" containerName="nova-manage" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827607 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c4e9807-4e89-49d0-bea0-732464e07ca2" containerName="nova-manage" Oct 09 10:25:22 crc kubenswrapper[4923]: E1009 10:25:22.827625 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340330a2-dbec-4301-a703-7ab727aad3d9" containerName="nova-metadata-metadata" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827632 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="340330a2-dbec-4301-a703-7ab727aad3d9" containerName="nova-metadata-metadata" Oct 09 10:25:22 crc kubenswrapper[4923]: E1009 10:25:22.827649 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="51933409-9af8-4cd7-8642-6d4f675068d7" containerName="dnsmasq-dns" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827656 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="51933409-9af8-4cd7-8642-6d4f675068d7" containerName="dnsmasq-dns" Oct 09 10:25:22 crc kubenswrapper[4923]: E1009 10:25:22.827676 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="340330a2-dbec-4301-a703-7ab727aad3d9" containerName="nova-metadata-log" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827685 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="340330a2-dbec-4301-a703-7ab727aad3d9" containerName="nova-metadata-log" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827912 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="340330a2-dbec-4301-a703-7ab727aad3d9" containerName="nova-metadata-log" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827931 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c4e9807-4e89-49d0-bea0-732464e07ca2" containerName="nova-manage" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827942 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="340330a2-dbec-4301-a703-7ab727aad3d9" containerName="nova-metadata-metadata" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.827953 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="51933409-9af8-4cd7-8642-6d4f675068d7" containerName="dnsmasq-dns" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.829075 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.830739 4923 scope.go:117] "RemoveContainer" containerID="59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63" Oct 09 10:25:22 crc kubenswrapper[4923]: E1009 10:25:22.831516 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63\": container with ID starting with 59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63 not found: ID does not exist" containerID="59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.831551 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63"} err="failed to get container status \"59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63\": rpc error: code = NotFound desc = could not find container \"59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63\": container with ID starting with 59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63 not found: ID does not exist" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.831574 4923 scope.go:117] "RemoveContainer" containerID="bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5" Oct 09 10:25:22 crc kubenswrapper[4923]: E1009 10:25:22.834212 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5\": container with ID starting with bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5 not found: ID does not exist" containerID="bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.834248 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5"} err="failed to get container status \"bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5\": rpc error: code = NotFound desc = could not find container \"bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5\": container with ID starting with bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5 not found: ID does not exist" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.834265 4923 scope.go:117] "RemoveContainer" containerID="59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.835396 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63"} err="failed to get container status \"59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63\": rpc error: code = NotFound desc = could not find container \"59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63\": container with ID starting with 59f1c4c503230cfe1ad01bef28355fc9bd0b972fa3d80e8c9f8bcdb01e4cab63 not found: ID does not exist" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.835551 4923 scope.go:117] "RemoveContainer" containerID="bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.836191 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.836500 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.837921 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5"} err="failed to get container status \"bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5\": rpc error: code = NotFound desc = could not find container \"bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5\": container with ID starting with bb5c7c34c3dd94464ac9ffd16ee38fe77c9c519a50db220a2b2f11ad8fb7fcc5 not found: ID does not exist" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.848610 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.998217 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-config-data\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.998320 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.998382 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rfkr\" (UniqueName: \"kubernetes.io/projected/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-kube-api-access-4rfkr\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.998434 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:22 crc kubenswrapper[4923]: I1009 10:25:22.998670 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-logs\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.101275 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.101408 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-logs\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.101609 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-config-data\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.101652 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.101724 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rfkr\" (UniqueName: \"kubernetes.io/projected/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-kube-api-access-4rfkr\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.102525 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-logs\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.107534 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.107953 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-config-data\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.111512 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.121190 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rfkr\" (UniqueName: \"kubernetes.io/projected/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-kube-api-access-4rfkr\") pod \"nova-metadata-0\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.298411 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.799204 4923 generic.go:334] "Generic (PLEG): container finished" podID="473dfb33-5dd7-4805-a57d-73fffe7e9bcd" containerID="4a90b2287c8ad383ed78f2913ee13f17eb04c81228bfb72a0e41822f619e498c" exitCode=0 Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.799652 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"473dfb33-5dd7-4805-a57d-73fffe7e9bcd","Type":"ContainerDied","Data":"4a90b2287c8ad383ed78f2913ee13f17eb04c81228bfb72a0e41822f619e498c"} Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.802979 4923 generic.go:334] "Generic (PLEG): container finished" podID="45dfedcc-9dbc-41e9-b3d8-e014b9409bf1" containerID="9d7fbf539527d18ed86c2341708e436a44ee2c29f3a95f9f99732acbacd9422b" exitCode=0 Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.803034 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-78qxk" event={"ID":"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1","Type":"ContainerDied","Data":"9d7fbf539527d18ed86c2341708e436a44ee2c29f3a95f9f99732acbacd9422b"} Oct 09 10:25:23 crc kubenswrapper[4923]: I1009 10:25:23.876828 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.078286 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.232222 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-combined-ca-bundle\") pod \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.232360 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-config-data\") pod \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.232403 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm5fl\" (UniqueName: \"kubernetes.io/projected/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-kube-api-access-xm5fl\") pod \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\" (UID: \"473dfb33-5dd7-4805-a57d-73fffe7e9bcd\") " Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.242662 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-kube-api-access-xm5fl" (OuterVolumeSpecName: "kube-api-access-xm5fl") pod "473dfb33-5dd7-4805-a57d-73fffe7e9bcd" (UID: "473dfb33-5dd7-4805-a57d-73fffe7e9bcd"). InnerVolumeSpecName "kube-api-access-xm5fl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.269554 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "473dfb33-5dd7-4805-a57d-73fffe7e9bcd" (UID: "473dfb33-5dd7-4805-a57d-73fffe7e9bcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.274473 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-config-data" (OuterVolumeSpecName: "config-data") pod "473dfb33-5dd7-4805-a57d-73fffe7e9bcd" (UID: "473dfb33-5dd7-4805-a57d-73fffe7e9bcd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.335288 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.335348 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.335363 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm5fl\" (UniqueName: \"kubernetes.io/projected/473dfb33-5dd7-4805-a57d-73fffe7e9bcd-kube-api-access-xm5fl\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.616708 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="340330a2-dbec-4301-a703-7ab727aad3d9" path="/var/lib/kubelet/pods/340330a2-dbec-4301-a703-7ab727aad3d9/volumes" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.820391 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"473dfb33-5dd7-4805-a57d-73fffe7e9bcd","Type":"ContainerDied","Data":"01c50b747d8369ed16534712ed820ecde0394e32706e6569dba9307feeaf1a32"} Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.820457 4923 scope.go:117] "RemoveContainer" containerID="4a90b2287c8ad383ed78f2913ee13f17eb04c81228bfb72a0e41822f619e498c" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.820602 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.830644 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40","Type":"ContainerStarted","Data":"bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2"} Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.830706 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40","Type":"ContainerStarted","Data":"5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080"} Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.830721 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40","Type":"ContainerStarted","Data":"a0446924177b9195a0098ac89936b8a92bdad0a0ac9460f0ac2be3c5e5039b1f"} Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.836541 4923 generic.go:334] "Generic (PLEG): container finished" podID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerID="837f0b14e2ef8b5458a707e89f723b9ff26d292f529528208fe839a167eab602" exitCode=0 Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.836702 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a111f6-7635-4a6d-9aab-fda23fcf4c54","Type":"ContainerDied","Data":"837f0b14e2ef8b5458a707e89f723b9ff26d292f529528208fe839a167eab602"} Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.859405 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.875833 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.899871 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.8998429310000002 podStartE2EDuration="2.899842931s" podCreationTimestamp="2025-10-09 10:25:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:24.88997316 +0000 UTC m=+1210.958154916" watchObservedRunningTime="2025-10-09 10:25:24.899842931 +0000 UTC m=+1210.968024687" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.911801 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:25:24 crc kubenswrapper[4923]: E1009 10:25:24.912540 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="473dfb33-5dd7-4805-a57d-73fffe7e9bcd" containerName="nova-scheduler-scheduler" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.912567 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="473dfb33-5dd7-4805-a57d-73fffe7e9bcd" containerName="nova-scheduler-scheduler" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.912798 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="473dfb33-5dd7-4805-a57d-73fffe7e9bcd" containerName="nova-scheduler-scheduler" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.913629 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.918414 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 09 10:25:24 crc kubenswrapper[4923]: I1009 10:25:24.923236 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.000547 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.050294 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-config-data\") pod \"nova-scheduler-0\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.050356 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.050442 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zrj8\" (UniqueName: \"kubernetes.io/projected/d216cf8d-f6a9-451a-8787-ff3708898dd6-kube-api-access-5zrj8\") pod \"nova-scheduler-0\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.152227 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-config-data\") pod \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.152377 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-logs\") pod \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.152605 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-combined-ca-bundle\") pod \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.152647 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czbwz\" (UniqueName: \"kubernetes.io/projected/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-kube-api-access-czbwz\") pod \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\" (UID: \"d3a111f6-7635-4a6d-9aab-fda23fcf4c54\") " Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.152930 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zrj8\" (UniqueName: \"kubernetes.io/projected/d216cf8d-f6a9-451a-8787-ff3708898dd6-kube-api-access-5zrj8\") pod \"nova-scheduler-0\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.153018 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-config-data\") pod \"nova-scheduler-0\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.153051 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.155106 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-logs" (OuterVolumeSpecName: "logs") pod "d3a111f6-7635-4a6d-9aab-fda23fcf4c54" (UID: "d3a111f6-7635-4a6d-9aab-fda23fcf4c54"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.161023 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-kube-api-access-czbwz" (OuterVolumeSpecName: "kube-api-access-czbwz") pod "d3a111f6-7635-4a6d-9aab-fda23fcf4c54" (UID: "d3a111f6-7635-4a6d-9aab-fda23fcf4c54"). InnerVolumeSpecName "kube-api-access-czbwz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.163608 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.166522 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-config-data\") pod \"nova-scheduler-0\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.177249 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zrj8\" (UniqueName: \"kubernetes.io/projected/d216cf8d-f6a9-451a-8787-ff3708898dd6-kube-api-access-5zrj8\") pod \"nova-scheduler-0\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.197745 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-config-data" (OuterVolumeSpecName: "config-data") pod "d3a111f6-7635-4a6d-9aab-fda23fcf4c54" (UID: "d3a111f6-7635-4a6d-9aab-fda23fcf4c54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.199846 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3a111f6-7635-4a6d-9aab-fda23fcf4c54" (UID: "d3a111f6-7635-4a6d-9aab-fda23fcf4c54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.223056 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.254675 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.254707 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.254717 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.254730 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czbwz\" (UniqueName: \"kubernetes.io/projected/d3a111f6-7635-4a6d-9aab-fda23fcf4c54-kube-api-access-czbwz\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.296594 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.356304 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-combined-ca-bundle\") pod \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.356688 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-config-data\") pod \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.357151 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-scripts\") pod \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.357361 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs4hs\" (UniqueName: \"kubernetes.io/projected/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-kube-api-access-rs4hs\") pod \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\" (UID: \"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1\") " Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.361919 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-scripts" (OuterVolumeSpecName: "scripts") pod "45dfedcc-9dbc-41e9-b3d8-e014b9409bf1" (UID: "45dfedcc-9dbc-41e9-b3d8-e014b9409bf1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.362403 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-kube-api-access-rs4hs" (OuterVolumeSpecName: "kube-api-access-rs4hs") pod "45dfedcc-9dbc-41e9-b3d8-e014b9409bf1" (UID: "45dfedcc-9dbc-41e9-b3d8-e014b9409bf1"). InnerVolumeSpecName "kube-api-access-rs4hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.391823 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-config-data" (OuterVolumeSpecName: "config-data") pod "45dfedcc-9dbc-41e9-b3d8-e014b9409bf1" (UID: "45dfedcc-9dbc-41e9-b3d8-e014b9409bf1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.394328 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "45dfedcc-9dbc-41e9-b3d8-e014b9409bf1" (UID: "45dfedcc-9dbc-41e9-b3d8-e014b9409bf1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.460592 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.460633 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.460646 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.460655 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs4hs\" (UniqueName: \"kubernetes.io/projected/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1-kube-api-access-rs4hs\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.783671 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:25:25 crc kubenswrapper[4923]: W1009 10:25:25.787928 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd216cf8d_f6a9_451a_8787_ff3708898dd6.slice/crio-1fa52b58814cb507731e9711ffc53469a5da4024577911c6d3a6cbacd3e75bca WatchSource:0}: Error finding container 1fa52b58814cb507731e9711ffc53469a5da4024577911c6d3a6cbacd3e75bca: Status 404 returned error can't find the container with id 1fa52b58814cb507731e9711ffc53469a5da4024577911c6d3a6cbacd3e75bca Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.849666 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d216cf8d-f6a9-451a-8787-ff3708898dd6","Type":"ContainerStarted","Data":"1fa52b58814cb507731e9711ffc53469a5da4024577911c6d3a6cbacd3e75bca"} Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.854134 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d3a111f6-7635-4a6d-9aab-fda23fcf4c54","Type":"ContainerDied","Data":"fe0c69f10457181d9a277223450b807c8568248e66d81cd7e84b149d093f0bb9"} Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.854165 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.854200 4923 scope.go:117] "RemoveContainer" containerID="837f0b14e2ef8b5458a707e89f723b9ff26d292f529528208fe839a167eab602" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.859439 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-78qxk" event={"ID":"45dfedcc-9dbc-41e9-b3d8-e014b9409bf1","Type":"ContainerDied","Data":"50bad4991bc744981af2567c75fc935ada0370130df332dc5c36a8b4b0f35973"} Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.859523 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50bad4991bc744981af2567c75fc935ada0370130df332dc5c36a8b4b0f35973" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.859713 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-78qxk" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.879274 4923 scope.go:117] "RemoveContainer" containerID="4ef03f06e925dbea844c1fb919b1c69d81d41ea701cce5de9b30356816c08f59" Oct 09 10:25:25 crc kubenswrapper[4923]: I1009 10:25:25.989777 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.012697 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.026537 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:26 crc kubenswrapper[4923]: E1009 10:25:26.027883 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-log" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.027904 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-log" Oct 09 10:25:26 crc kubenswrapper[4923]: E1009 10:25:26.027915 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-api" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.027922 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-api" Oct 09 10:25:26 crc kubenswrapper[4923]: E1009 10:25:26.027934 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45dfedcc-9dbc-41e9-b3d8-e014b9409bf1" containerName="nova-cell1-conductor-db-sync" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.027940 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="45dfedcc-9dbc-41e9-b3d8-e014b9409bf1" containerName="nova-cell1-conductor-db-sync" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.029781 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="45dfedcc-9dbc-41e9-b3d8-e014b9409bf1" containerName="nova-cell1-conductor-db-sync" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.029806 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-api" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.029831 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" containerName="nova-api-log" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.031117 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.035667 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.040833 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.042664 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.049677 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.068397 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.079657 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.184405 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37e4f72b-a6e4-4f99-97f0-d367ce91b533-logs\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.184484 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c043f4-dce1-4779-8b34-6b988207c784-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"99c043f4-dce1-4779-8b34-6b988207c784\") " pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.184649 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c043f4-dce1-4779-8b34-6b988207c784-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"99c043f4-dce1-4779-8b34-6b988207c784\") " pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.184801 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64975\" (UniqueName: \"kubernetes.io/projected/37e4f72b-a6e4-4f99-97f0-d367ce91b533-kube-api-access-64975\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.185043 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-config-data\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.185085 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2r85l\" (UniqueName: \"kubernetes.io/projected/99c043f4-dce1-4779-8b34-6b988207c784-kube-api-access-2r85l\") pod \"nova-cell1-conductor-0\" (UID: \"99c043f4-dce1-4779-8b34-6b988207c784\") " pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.185123 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.290092 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-config-data\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.290607 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2r85l\" (UniqueName: \"kubernetes.io/projected/99c043f4-dce1-4779-8b34-6b988207c784-kube-api-access-2r85l\") pod \"nova-cell1-conductor-0\" (UID: \"99c043f4-dce1-4779-8b34-6b988207c784\") " pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.290636 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.290725 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37e4f72b-a6e4-4f99-97f0-d367ce91b533-logs\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.290784 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c043f4-dce1-4779-8b34-6b988207c784-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"99c043f4-dce1-4779-8b34-6b988207c784\") " pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.290825 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c043f4-dce1-4779-8b34-6b988207c784-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"99c043f4-dce1-4779-8b34-6b988207c784\") " pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.290883 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64975\" (UniqueName: \"kubernetes.io/projected/37e4f72b-a6e4-4f99-97f0-d367ce91b533-kube-api-access-64975\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.291455 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37e4f72b-a6e4-4f99-97f0-d367ce91b533-logs\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.296868 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.298405 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-config-data\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.299638 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99c043f4-dce1-4779-8b34-6b988207c784-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"99c043f4-dce1-4779-8b34-6b988207c784\") " pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.302432 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c043f4-dce1-4779-8b34-6b988207c784-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"99c043f4-dce1-4779-8b34-6b988207c784\") " pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.317770 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64975\" (UniqueName: \"kubernetes.io/projected/37e4f72b-a6e4-4f99-97f0-d367ce91b533-kube-api-access-64975\") pod \"nova-api-0\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.317893 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2r85l\" (UniqueName: \"kubernetes.io/projected/99c043f4-dce1-4779-8b34-6b988207c784-kube-api-access-2r85l\") pod \"nova-cell1-conductor-0\" (UID: \"99c043f4-dce1-4779-8b34-6b988207c784\") " pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.365551 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.376872 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.616671 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="473dfb33-5dd7-4805-a57d-73fffe7e9bcd" path="/var/lib/kubelet/pods/473dfb33-5dd7-4805-a57d-73fffe7e9bcd/volumes" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.617438 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3a111f6-7635-4a6d-9aab-fda23fcf4c54" path="/var/lib/kubelet/pods/d3a111f6-7635-4a6d-9aab-fda23fcf4c54/volumes" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.876581 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d216cf8d-f6a9-451a-8787-ff3708898dd6","Type":"ContainerStarted","Data":"6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec"} Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.916162 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.91613182 podStartE2EDuration="2.91613182s" podCreationTimestamp="2025-10-09 10:25:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:26.911642717 +0000 UTC m=+1212.979824473" watchObservedRunningTime="2025-10-09 10:25:26.91613182 +0000 UTC m=+1212.984313566" Oct 09 10:25:26 crc kubenswrapper[4923]: I1009 10:25:26.948681 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:27 crc kubenswrapper[4923]: I1009 10:25:27.021609 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 09 10:25:27 crc kubenswrapper[4923]: I1009 10:25:27.897388 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"99c043f4-dce1-4779-8b34-6b988207c784","Type":"ContainerStarted","Data":"22cddb5e1104d1f4c3ac7d31ec1f64b61d169222c381bf31c8c5e88cb5f68a71"} Oct 09 10:25:27 crc kubenswrapper[4923]: I1009 10:25:27.897992 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:27 crc kubenswrapper[4923]: I1009 10:25:27.898012 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"99c043f4-dce1-4779-8b34-6b988207c784","Type":"ContainerStarted","Data":"7e0a9808d5dfbb5e4f3ac6e5ac0ffc18a63e8dfecd47ebaeff1929f9d126fb0d"} Oct 09 10:25:27 crc kubenswrapper[4923]: I1009 10:25:27.900387 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37e4f72b-a6e4-4f99-97f0-d367ce91b533","Type":"ContainerStarted","Data":"69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0"} Oct 09 10:25:27 crc kubenswrapper[4923]: I1009 10:25:27.900550 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37e4f72b-a6e4-4f99-97f0-d367ce91b533","Type":"ContainerStarted","Data":"6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65"} Oct 09 10:25:27 crc kubenswrapper[4923]: I1009 10:25:27.900635 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37e4f72b-a6e4-4f99-97f0-d367ce91b533","Type":"ContainerStarted","Data":"43581a5132da753321b8ccd0869eaa746c2b46e1ab7419c26f092c3581e34caa"} Oct 09 10:25:27 crc kubenswrapper[4923]: I1009 10:25:27.952568 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.9525352160000002 podStartE2EDuration="2.952535216s" podCreationTimestamp="2025-10-09 10:25:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:27.945113321 +0000 UTC m=+1214.013295077" watchObservedRunningTime="2025-10-09 10:25:27.952535216 +0000 UTC m=+1214.020716972" Oct 09 10:25:27 crc kubenswrapper[4923]: I1009 10:25:27.954894 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.95488014 podStartE2EDuration="2.95488014s" podCreationTimestamp="2025-10-09 10:25:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:27.926439709 +0000 UTC m=+1213.994621465" watchObservedRunningTime="2025-10-09 10:25:27.95488014 +0000 UTC m=+1214.023061896" Oct 09 10:25:28 crc kubenswrapper[4923]: I1009 10:25:28.299343 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 10:25:28 crc kubenswrapper[4923]: I1009 10:25:28.300300 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 10:25:28 crc kubenswrapper[4923]: I1009 10:25:28.695014 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 09 10:25:30 crc kubenswrapper[4923]: I1009 10:25:30.299846 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 09 10:25:33 crc kubenswrapper[4923]: I1009 10:25:33.299132 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 09 10:25:33 crc kubenswrapper[4923]: I1009 10:25:33.299783 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 09 10:25:34 crc kubenswrapper[4923]: I1009 10:25:34.311983 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 10:25:34 crc kubenswrapper[4923]: I1009 10:25:34.312074 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 10:25:35 crc kubenswrapper[4923]: I1009 10:25:35.296953 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 09 10:25:35 crc kubenswrapper[4923]: I1009 10:25:35.334983 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 09 10:25:36 crc kubenswrapper[4923]: I1009 10:25:36.034910 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 09 10:25:36 crc kubenswrapper[4923]: I1009 10:25:36.366457 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 10:25:36 crc kubenswrapper[4923]: I1009 10:25:36.366533 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 10:25:36 crc kubenswrapper[4923]: I1009 10:25:36.410332 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 09 10:25:37 crc kubenswrapper[4923]: I1009 10:25:37.407096 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 10:25:37 crc kubenswrapper[4923]: I1009 10:25:37.448202 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.185:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.043894 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.061349 4923 generic.go:334] "Generic (PLEG): container finished" podID="ed966e00-bc66-4579-b9b1-e16c7dd82854" containerID="80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d" exitCode=137 Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.061429 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.061459 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed966e00-bc66-4579-b9b1-e16c7dd82854","Type":"ContainerDied","Data":"80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d"} Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.061498 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ed966e00-bc66-4579-b9b1-e16c7dd82854","Type":"ContainerDied","Data":"997c76ee6a3cb56c92e3a94cd2e3702a72b6ed923a52ff05ce2bc239ba17d374"} Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.061521 4923 scope.go:117] "RemoveContainer" containerID="80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.104869 4923 scope.go:117] "RemoveContainer" containerID="80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d" Oct 09 10:25:43 crc kubenswrapper[4923]: E1009 10:25:43.105607 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d\": container with ID starting with 80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d not found: ID does not exist" containerID="80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.105670 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d"} err="failed to get container status \"80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d\": rpc error: code = NotFound desc = could not find container \"80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d\": container with ID starting with 80b5ebb270a14d01597480b9a26cbc6747160dd4a175ff87d6528d86cd03b67d not found: ID does not exist" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.193570 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rddp\" (UniqueName: \"kubernetes.io/projected/ed966e00-bc66-4579-b9b1-e16c7dd82854-kube-api-access-9rddp\") pod \"ed966e00-bc66-4579-b9b1-e16c7dd82854\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.193702 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-combined-ca-bundle\") pod \"ed966e00-bc66-4579-b9b1-e16c7dd82854\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.193829 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-config-data\") pod \"ed966e00-bc66-4579-b9b1-e16c7dd82854\" (UID: \"ed966e00-bc66-4579-b9b1-e16c7dd82854\") " Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.204397 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed966e00-bc66-4579-b9b1-e16c7dd82854-kube-api-access-9rddp" (OuterVolumeSpecName: "kube-api-access-9rddp") pod "ed966e00-bc66-4579-b9b1-e16c7dd82854" (UID: "ed966e00-bc66-4579-b9b1-e16c7dd82854"). InnerVolumeSpecName "kube-api-access-9rddp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.226895 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ed966e00-bc66-4579-b9b1-e16c7dd82854" (UID: "ed966e00-bc66-4579-b9b1-e16c7dd82854"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.230422 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-config-data" (OuterVolumeSpecName: "config-data") pod "ed966e00-bc66-4579-b9b1-e16c7dd82854" (UID: "ed966e00-bc66-4579-b9b1-e16c7dd82854"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.296856 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.297225 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rddp\" (UniqueName: \"kubernetes.io/projected/ed966e00-bc66-4579-b9b1-e16c7dd82854-kube-api-access-9rddp\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.297312 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed966e00-bc66-4579-b9b1-e16c7dd82854-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.310096 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.310445 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.314598 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.404409 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.415976 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.428188 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 10:25:43 crc kubenswrapper[4923]: E1009 10:25:43.428872 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed966e00-bc66-4579-b9b1-e16c7dd82854" containerName="nova-cell1-novncproxy-novncproxy" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.428913 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed966e00-bc66-4579-b9b1-e16c7dd82854" containerName="nova-cell1-novncproxy-novncproxy" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.429191 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed966e00-bc66-4579-b9b1-e16c7dd82854" containerName="nova-cell1-novncproxy-novncproxy" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.430135 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.439425 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.440148 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.440488 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.441340 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.501369 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.501779 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.501896 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.501984 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.502095 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9h7l\" (UniqueName: \"kubernetes.io/projected/8b60a796-c363-488e-9f05-1dedbf605ab5-kube-api-access-n9h7l\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.604286 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.604395 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.604451 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.604499 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.604568 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9h7l\" (UniqueName: \"kubernetes.io/projected/8b60a796-c363-488e-9f05-1dedbf605ab5-kube-api-access-n9h7l\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.609526 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.609734 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.610028 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.610324 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8b60a796-c363-488e-9f05-1dedbf605ab5-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.622980 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9h7l\" (UniqueName: \"kubernetes.io/projected/8b60a796-c363-488e-9f05-1dedbf605ab5-kube-api-access-n9h7l\") pod \"nova-cell1-novncproxy-0\" (UID: \"8b60a796-c363-488e-9f05-1dedbf605ab5\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:43 crc kubenswrapper[4923]: I1009 10:25:43.752230 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:44 crc kubenswrapper[4923]: I1009 10:25:44.118939 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 09 10:25:44 crc kubenswrapper[4923]: I1009 10:25:44.320077 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 10:25:44 crc kubenswrapper[4923]: W1009 10:25:44.324012 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b60a796_c363_488e_9f05_1dedbf605ab5.slice/crio-a521fd818113650c4455c4b5c367c0bc7c1d4e933ae2bd72987da229e8697def WatchSource:0}: Error finding container a521fd818113650c4455c4b5c367c0bc7c1d4e933ae2bd72987da229e8697def: Status 404 returned error can't find the container with id a521fd818113650c4455c4b5c367c0bc7c1d4e933ae2bd72987da229e8697def Oct 09 10:25:44 crc kubenswrapper[4923]: I1009 10:25:44.610683 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed966e00-bc66-4579-b9b1-e16c7dd82854" path="/var/lib/kubelet/pods/ed966e00-bc66-4579-b9b1-e16c7dd82854/volumes" Oct 09 10:25:45 crc kubenswrapper[4923]: I1009 10:25:45.105946 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8b60a796-c363-488e-9f05-1dedbf605ab5","Type":"ContainerStarted","Data":"3cb5f26da5a663d295f49cc9b187100bc43c6b515bc4a530903ee2c948a1b954"} Oct 09 10:25:45 crc kubenswrapper[4923]: I1009 10:25:45.106033 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"8b60a796-c363-488e-9f05-1dedbf605ab5","Type":"ContainerStarted","Data":"a521fd818113650c4455c4b5c367c0bc7c1d4e933ae2bd72987da229e8697def"} Oct 09 10:25:45 crc kubenswrapper[4923]: I1009 10:25:45.131209 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.131180939 podStartE2EDuration="2.131180939s" podCreationTimestamp="2025-10-09 10:25:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:45.128442804 +0000 UTC m=+1231.196624580" watchObservedRunningTime="2025-10-09 10:25:45.131180939 +0000 UTC m=+1231.199362695" Oct 09 10:25:46 crc kubenswrapper[4923]: I1009 10:25:46.370916 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 09 10:25:46 crc kubenswrapper[4923]: I1009 10:25:46.374307 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 09 10:25:46 crc kubenswrapper[4923]: I1009 10:25:46.375609 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 09 10:25:46 crc kubenswrapper[4923]: I1009 10:25:46.386437 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.124743 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.129723 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.347330 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869677f947-lv65r"] Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.351847 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.382434 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869677f947-lv65r"] Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.426733 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-nb\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.426813 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-dns-svc\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.426937 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-config\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.427313 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-sb\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.427382 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z7sv\" (UniqueName: \"kubernetes.io/projected/462961cd-6180-4e0a-84b6-cae4c483c869-kube-api-access-7z7sv\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.528999 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-sb\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.529085 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z7sv\" (UniqueName: \"kubernetes.io/projected/462961cd-6180-4e0a-84b6-cae4c483c869-kube-api-access-7z7sv\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.529207 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-nb\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.529244 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-dns-svc\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.529329 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-config\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.531150 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-dns-svc\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.531313 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-config\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.531368 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-sb\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.531955 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-nb\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.552294 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z7sv\" (UniqueName: \"kubernetes.io/projected/462961cd-6180-4e0a-84b6-cae4c483c869-kube-api-access-7z7sv\") pod \"dnsmasq-dns-869677f947-lv65r\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:47 crc kubenswrapper[4923]: I1009 10:25:47.684094 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:48 crc kubenswrapper[4923]: I1009 10:25:48.305211 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869677f947-lv65r"] Oct 09 10:25:48 crc kubenswrapper[4923]: I1009 10:25:48.752742 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:49 crc kubenswrapper[4923]: I1009 10:25:49.150213 4923 generic.go:334] "Generic (PLEG): container finished" podID="462961cd-6180-4e0a-84b6-cae4c483c869" containerID="10448a2d329eefb5a66ecf2d245104d0fe16f581201818ca1ddc07314ba7554c" exitCode=0 Oct 09 10:25:49 crc kubenswrapper[4923]: I1009 10:25:49.151888 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-lv65r" event={"ID":"462961cd-6180-4e0a-84b6-cae4c483c869","Type":"ContainerDied","Data":"10448a2d329eefb5a66ecf2d245104d0fe16f581201818ca1ddc07314ba7554c"} Oct 09 10:25:49 crc kubenswrapper[4923]: I1009 10:25:49.151934 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-lv65r" event={"ID":"462961cd-6180-4e0a-84b6-cae4c483c869","Type":"ContainerStarted","Data":"aba00e6aa229b8364cea92cd6ba987eb92800d6ee5be5d3809f631ac2278b1b0"} Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.080765 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.081545 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="ceilometer-central-agent" containerID="cri-o://95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625" gracePeriod=30 Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.081621 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="proxy-httpd" containerID="cri-o://0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859" gracePeriod=30 Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.081720 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="sg-core" containerID="cri-o://a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d" gracePeriod=30 Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.081814 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="ceilometer-notification-agent" containerID="cri-o://21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d" gracePeriod=30 Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.163018 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-lv65r" event={"ID":"462961cd-6180-4e0a-84b6-cae4c483c869","Type":"ContainerStarted","Data":"472feb9272ddee5f54959bd6693c5376a5402b3f2a8b8d70965328c54d68b419"} Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.163260 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.186324 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-869677f947-lv65r" podStartSLOduration=3.18629217 podStartE2EDuration="3.18629217s" podCreationTimestamp="2025-10-09 10:25:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:50.183708999 +0000 UTC m=+1236.251890765" watchObservedRunningTime="2025-10-09 10:25:50.18629217 +0000 UTC m=+1236.254473936" Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.254018 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.254300 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-log" containerID="cri-o://6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65" gracePeriod=30 Oct 09 10:25:50 crc kubenswrapper[4923]: I1009 10:25:50.254420 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-api" containerID="cri-o://69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0" gracePeriod=30 Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.177631 4923 generic.go:334] "Generic (PLEG): container finished" podID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerID="0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859" exitCode=0 Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.178905 4923 generic.go:334] "Generic (PLEG): container finished" podID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerID="a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d" exitCode=2 Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.178984 4923 generic.go:334] "Generic (PLEG): container finished" podID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerID="95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625" exitCode=0 Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.177814 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerDied","Data":"0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859"} Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.179230 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerDied","Data":"a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d"} Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.179303 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerDied","Data":"95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625"} Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.187208 4923 generic.go:334] "Generic (PLEG): container finished" podID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerID="6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65" exitCode=143 Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.188193 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37e4f72b-a6e4-4f99-97f0-d367ce91b533","Type":"ContainerDied","Data":"6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65"} Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.829588 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.833366 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-run-httpd\") pod \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.833421 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-log-httpd\") pod \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.833461 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-ceilometer-tls-certs\") pod \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.833492 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-combined-ca-bundle\") pod \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.833522 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxzhv\" (UniqueName: \"kubernetes.io/projected/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-kube-api-access-zxzhv\") pod \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.833589 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-config-data\") pod \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.833659 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-sg-core-conf-yaml\") pod \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.833703 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-scripts\") pod \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\" (UID: \"c3214d04-5a88-4f5a-bb3a-9c3d172981ea\") " Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.835189 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c3214d04-5a88-4f5a-bb3a-9c3d172981ea" (UID: "c3214d04-5a88-4f5a-bb3a-9c3d172981ea"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.835990 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c3214d04-5a88-4f5a-bb3a-9c3d172981ea" (UID: "c3214d04-5a88-4f5a-bb3a-9c3d172981ea"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.842918 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-scripts" (OuterVolumeSpecName: "scripts") pod "c3214d04-5a88-4f5a-bb3a-9c3d172981ea" (UID: "c3214d04-5a88-4f5a-bb3a-9c3d172981ea"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.845353 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-kube-api-access-zxzhv" (OuterVolumeSpecName: "kube-api-access-zxzhv") pod "c3214d04-5a88-4f5a-bb3a-9c3d172981ea" (UID: "c3214d04-5a88-4f5a-bb3a-9c3d172981ea"). InnerVolumeSpecName "kube-api-access-zxzhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.915014 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c3214d04-5a88-4f5a-bb3a-9c3d172981ea" (UID: "c3214d04-5a88-4f5a-bb3a-9c3d172981ea"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.937799 4923 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.937858 4923 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.937877 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxzhv\" (UniqueName: \"kubernetes.io/projected/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-kube-api-access-zxzhv\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.937892 4923 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.937907 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.977368 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c3214d04-5a88-4f5a-bb3a-9c3d172981ea" (UID: "c3214d04-5a88-4f5a-bb3a-9c3d172981ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:51 crc kubenswrapper[4923]: I1009 10:25:51.980670 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "c3214d04-5a88-4f5a-bb3a-9c3d172981ea" (UID: "c3214d04-5a88-4f5a-bb3a-9c3d172981ea"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.021875 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-config-data" (OuterVolumeSpecName: "config-data") pod "c3214d04-5a88-4f5a-bb3a-9c3d172981ea" (UID: "c3214d04-5a88-4f5a-bb3a-9c3d172981ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.039972 4923 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.040023 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.040040 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c3214d04-5a88-4f5a-bb3a-9c3d172981ea-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.200998 4923 generic.go:334] "Generic (PLEG): container finished" podID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerID="21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d" exitCode=0 Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.201064 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerDied","Data":"21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d"} Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.201105 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c3214d04-5a88-4f5a-bb3a-9c3d172981ea","Type":"ContainerDied","Data":"67119f01b6905ea3287bc7bddc5652c242bbc14ba90c4dd3ddd3ee630231339d"} Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.201120 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.201133 4923 scope.go:117] "RemoveContainer" containerID="0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.231841 4923 scope.go:117] "RemoveContainer" containerID="a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.260708 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.293125 4923 scope.go:117] "RemoveContainer" containerID="21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.299990 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.333808 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:25:52 crc kubenswrapper[4923]: E1009 10:25:52.334380 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="ceilometer-notification-agent" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.334407 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="ceilometer-notification-agent" Oct 09 10:25:52 crc kubenswrapper[4923]: E1009 10:25:52.334429 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="ceilometer-central-agent" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.334439 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="ceilometer-central-agent" Oct 09 10:25:52 crc kubenswrapper[4923]: E1009 10:25:52.334461 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="sg-core" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.334471 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="sg-core" Oct 09 10:25:52 crc kubenswrapper[4923]: E1009 10:25:52.334489 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="proxy-httpd" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.334496 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="proxy-httpd" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.334705 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="ceilometer-central-agent" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.334720 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="sg-core" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.334730 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="ceilometer-notification-agent" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.334747 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" containerName="proxy-httpd" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.336981 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.345084 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.345326 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.345487 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.362603 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.378015 4923 scope.go:117] "RemoveContainer" containerID="95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.380232 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.380514 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnwc6\" (UniqueName: \"kubernetes.io/projected/38dde525-347c-4d29-bbd7-20a1e7712193-kube-api-access-dnwc6\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.380582 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-scripts\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.380775 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.380812 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-log-httpd\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.380866 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-config-data\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.380893 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.380951 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-run-httpd\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.415378 4923 scope.go:117] "RemoveContainer" containerID="0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859" Oct 09 10:25:52 crc kubenswrapper[4923]: E1009 10:25:52.416346 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859\": container with ID starting with 0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859 not found: ID does not exist" containerID="0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.416473 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859"} err="failed to get container status \"0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859\": rpc error: code = NotFound desc = could not find container \"0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859\": container with ID starting with 0c0728fddbfd15dfcc8c9c71ba8aba537505046372b615ff541b2b22ea57e859 not found: ID does not exist" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.416517 4923 scope.go:117] "RemoveContainer" containerID="a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d" Oct 09 10:25:52 crc kubenswrapper[4923]: E1009 10:25:52.417282 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d\": container with ID starting with a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d not found: ID does not exist" containerID="a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.417325 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d"} err="failed to get container status \"a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d\": rpc error: code = NotFound desc = could not find container \"a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d\": container with ID starting with a954eccb1a0bb2c6944be9758e3bd48664fd5a392d05427d199ad481ed7fc38d not found: ID does not exist" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.417355 4923 scope.go:117] "RemoveContainer" containerID="21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d" Oct 09 10:25:52 crc kubenswrapper[4923]: E1009 10:25:52.417910 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d\": container with ID starting with 21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d not found: ID does not exist" containerID="21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.417942 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d"} err="failed to get container status \"21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d\": rpc error: code = NotFound desc = could not find container \"21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d\": container with ID starting with 21cbc5d3b320daf4afe5b223eae5b7ca8218e4c2fb27e29f57b5a66ea4eb8e5d not found: ID does not exist" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.417960 4923 scope.go:117] "RemoveContainer" containerID="95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625" Oct 09 10:25:52 crc kubenswrapper[4923]: E1009 10:25:52.418377 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625\": container with ID starting with 95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625 not found: ID does not exist" containerID="95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.418418 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625"} err="failed to get container status \"95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625\": rpc error: code = NotFound desc = could not find container \"95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625\": container with ID starting with 95f6597eeb15d0915bac314604cea6aaa7dd217adab2ca57e8d5e4247a806625 not found: ID does not exist" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.483445 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnwc6\" (UniqueName: \"kubernetes.io/projected/38dde525-347c-4d29-bbd7-20a1e7712193-kube-api-access-dnwc6\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.483520 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-scripts\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.483594 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.483616 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-log-httpd\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.483647 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-config-data\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.483672 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.483707 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-run-httpd\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.483850 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.485518 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-log-httpd\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.486284 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-run-httpd\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.489712 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.490330 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.491698 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.491949 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-config-data\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.494121 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-scripts\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.505947 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnwc6\" (UniqueName: \"kubernetes.io/projected/38dde525-347c-4d29-bbd7-20a1e7712193-kube-api-access-dnwc6\") pod \"ceilometer-0\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.614551 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3214d04-5a88-4f5a-bb3a-9c3d172981ea" path="/var/lib/kubelet/pods/c3214d04-5a88-4f5a-bb3a-9c3d172981ea/volumes" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.667737 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:25:52 crc kubenswrapper[4923]: I1009 10:25:52.723836 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:25:53 crc kubenswrapper[4923]: I1009 10:25:53.214031 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:25:53 crc kubenswrapper[4923]: W1009 10:25:53.219621 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38dde525_347c_4d29_bbd7_20a1e7712193.slice/crio-7b31e1526307b9273759fa0c24cbee3269a59fa7bdcf529bbb9c646abe668748 WatchSource:0}: Error finding container 7b31e1526307b9273759fa0c24cbee3269a59fa7bdcf529bbb9c646abe668748: Status 404 returned error can't find the container with id 7b31e1526307b9273759fa0c24cbee3269a59fa7bdcf529bbb9c646abe668748 Oct 09 10:25:53 crc kubenswrapper[4923]: I1009 10:25:53.227357 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 10:25:53 crc kubenswrapper[4923]: I1009 10:25:53.752872 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:53 crc kubenswrapper[4923]: I1009 10:25:53.781878 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:53 crc kubenswrapper[4923]: I1009 10:25:53.995976 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.125388 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-config-data\") pod \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.125495 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64975\" (UniqueName: \"kubernetes.io/projected/37e4f72b-a6e4-4f99-97f0-d367ce91b533-kube-api-access-64975\") pod \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.125563 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37e4f72b-a6e4-4f99-97f0-d367ce91b533-logs\") pod \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.125745 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-combined-ca-bundle\") pod \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\" (UID: \"37e4f72b-a6e4-4f99-97f0-d367ce91b533\") " Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.129335 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37e4f72b-a6e4-4f99-97f0-d367ce91b533-logs" (OuterVolumeSpecName: "logs") pod "37e4f72b-a6e4-4f99-97f0-d367ce91b533" (UID: "37e4f72b-a6e4-4f99-97f0-d367ce91b533"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.144126 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37e4f72b-a6e4-4f99-97f0-d367ce91b533-kube-api-access-64975" (OuterVolumeSpecName: "kube-api-access-64975") pod "37e4f72b-a6e4-4f99-97f0-d367ce91b533" (UID: "37e4f72b-a6e4-4f99-97f0-d367ce91b533"). InnerVolumeSpecName "kube-api-access-64975". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.173554 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37e4f72b-a6e4-4f99-97f0-d367ce91b533" (UID: "37e4f72b-a6e4-4f99-97f0-d367ce91b533"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.216860 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-config-data" (OuterVolumeSpecName: "config-data") pod "37e4f72b-a6e4-4f99-97f0-d367ce91b533" (UID: "37e4f72b-a6e4-4f99-97f0-d367ce91b533"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.228016 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64975\" (UniqueName: \"kubernetes.io/projected/37e4f72b-a6e4-4f99-97f0-d367ce91b533-kube-api-access-64975\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.228053 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37e4f72b-a6e4-4f99-97f0-d367ce91b533-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.228069 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.228082 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37e4f72b-a6e4-4f99-97f0-d367ce91b533-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.228884 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerStarted","Data":"7b31e1526307b9273759fa0c24cbee3269a59fa7bdcf529bbb9c646abe668748"} Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.231271 4923 generic.go:334] "Generic (PLEG): container finished" podID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerID="69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0" exitCode=0 Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.232092 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.232164 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37e4f72b-a6e4-4f99-97f0-d367ce91b533","Type":"ContainerDied","Data":"69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0"} Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.232255 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"37e4f72b-a6e4-4f99-97f0-d367ce91b533","Type":"ContainerDied","Data":"43581a5132da753321b8ccd0869eaa746c2b46e1ab7419c26f092c3581e34caa"} Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.232290 4923 scope.go:117] "RemoveContainer" containerID="69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.260148 4923 scope.go:117] "RemoveContainer" containerID="6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.266549 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.297993 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.309270 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.317524 4923 scope.go:117] "RemoveContainer" containerID="69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0" Oct 09 10:25:54 crc kubenswrapper[4923]: E1009 10:25:54.330815 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0\": container with ID starting with 69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0 not found: ID does not exist" containerID="69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.330882 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0"} err="failed to get container status \"69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0\": rpc error: code = NotFound desc = could not find container \"69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0\": container with ID starting with 69a022b5cfb20f3dca473529a72e077b6806f598f58734dfac942574f8d9eda0 not found: ID does not exist" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.330929 4923 scope.go:117] "RemoveContainer" containerID="6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65" Oct 09 10:25:54 crc kubenswrapper[4923]: E1009 10:25:54.331308 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65\": container with ID starting with 6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65 not found: ID does not exist" containerID="6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.331333 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65"} err="failed to get container status \"6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65\": rpc error: code = NotFound desc = could not find container \"6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65\": container with ID starting with 6792429b9b7a435034090727c91c4e2e447ab6830e115b5c79a5b7ba35efdd65 not found: ID does not exist" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.369930 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:54 crc kubenswrapper[4923]: E1009 10:25:54.376243 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-api" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.376278 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-api" Oct 09 10:25:54 crc kubenswrapper[4923]: E1009 10:25:54.376292 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-log" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.376302 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-log" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.376502 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-log" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.376535 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" containerName="nova-api-api" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.377615 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.381292 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.381660 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.381793 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.407938 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.535874 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-public-tls-certs\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.535994 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.536366 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-config-data\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.536457 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69gs6\" (UniqueName: \"kubernetes.io/projected/5eb00329-d879-4dbe-b9ca-25491082df45-kube-api-access-69gs6\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.536497 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.536578 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eb00329-d879-4dbe-b9ca-25491082df45-logs\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.545923 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-b2zqk"] Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.547474 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.551268 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.551713 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.556557 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-b2zqk"] Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.619836 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37e4f72b-a6e4-4f99-97f0-d367ce91b533" path="/var/lib/kubelet/pods/37e4f72b-a6e4-4f99-97f0-d367ce91b533/volumes" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638378 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-config-data\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638468 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-config-data\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638510 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-scripts\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638563 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69gs6\" (UniqueName: \"kubernetes.io/projected/5eb00329-d879-4dbe-b9ca-25491082df45-kube-api-access-69gs6\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638599 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638624 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638650 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l66v4\" (UniqueName: \"kubernetes.io/projected/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-kube-api-access-l66v4\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638701 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eb00329-d879-4dbe-b9ca-25491082df45-logs\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638911 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-public-tls-certs\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.638989 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.640069 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eb00329-d879-4dbe-b9ca-25491082df45-logs\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.644713 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-internal-tls-certs\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.644731 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.645472 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-config-data\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.652598 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-public-tls-certs\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.665358 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69gs6\" (UniqueName: \"kubernetes.io/projected/5eb00329-d879-4dbe-b9ca-25491082df45-kube-api-access-69gs6\") pod \"nova-api-0\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.740976 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-config-data\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.741053 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-scripts\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.741095 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.741117 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l66v4\" (UniqueName: \"kubernetes.io/projected/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-kube-api-access-l66v4\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.745253 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-scripts\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.745370 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.746773 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-config-data\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.762483 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l66v4\" (UniqueName: \"kubernetes.io/projected/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-kube-api-access-l66v4\") pod \"nova-cell1-cell-mapping-b2zqk\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.764574 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:25:54 crc kubenswrapper[4923]: I1009 10:25:54.878825 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:25:55 crc kubenswrapper[4923]: I1009 10:25:55.252964 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerStarted","Data":"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660"} Oct 09 10:25:55 crc kubenswrapper[4923]: I1009 10:25:55.296589 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:25:55 crc kubenswrapper[4923]: W1009 10:25:55.308516 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5eb00329_d879_4dbe_b9ca_25491082df45.slice/crio-2ff31e1c716d0703ed16cb9d742b3a49620048b700e884d642cb1af13c0687d6 WatchSource:0}: Error finding container 2ff31e1c716d0703ed16cb9d742b3a49620048b700e884d642cb1af13c0687d6: Status 404 returned error can't find the container with id 2ff31e1c716d0703ed16cb9d742b3a49620048b700e884d642cb1af13c0687d6 Oct 09 10:25:55 crc kubenswrapper[4923]: I1009 10:25:55.436982 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-b2zqk"] Oct 09 10:25:56 crc kubenswrapper[4923]: I1009 10:25:56.271303 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eb00329-d879-4dbe-b9ca-25491082df45","Type":"ContainerStarted","Data":"2d366a97b0924e8d3eeb3478d7fc9dfeec921a6c51bda9d8b02cbb7e579c03a4"} Oct 09 10:25:56 crc kubenswrapper[4923]: I1009 10:25:56.271913 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eb00329-d879-4dbe-b9ca-25491082df45","Type":"ContainerStarted","Data":"df7c565a2ba98b9c9a71d71dcec291fecef54730336abc5d1f3be7d4b2eb3fa0"} Oct 09 10:25:56 crc kubenswrapper[4923]: I1009 10:25:56.271931 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eb00329-d879-4dbe-b9ca-25491082df45","Type":"ContainerStarted","Data":"2ff31e1c716d0703ed16cb9d742b3a49620048b700e884d642cb1af13c0687d6"} Oct 09 10:25:56 crc kubenswrapper[4923]: I1009 10:25:56.273227 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b2zqk" event={"ID":"daf12e8f-fbe9-44dd-aff6-c8033ae56d45","Type":"ContainerStarted","Data":"48f3f03b0b0e33fb4cb6af6c4d9de2cdc3336e631ffa235c713c4136c63a2637"} Oct 09 10:25:56 crc kubenswrapper[4923]: I1009 10:25:56.273289 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b2zqk" event={"ID":"daf12e8f-fbe9-44dd-aff6-c8033ae56d45","Type":"ContainerStarted","Data":"3194c9bac6ec3175f98ebca7f847b8d826df0946842d3a1c7513e8f14ad6beeb"} Oct 09 10:25:56 crc kubenswrapper[4923]: I1009 10:25:56.275580 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerStarted","Data":"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128"} Oct 09 10:25:56 crc kubenswrapper[4923]: I1009 10:25:56.302506 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.3024734159999998 podStartE2EDuration="2.302473416s" podCreationTimestamp="2025-10-09 10:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:56.293556532 +0000 UTC m=+1242.361738298" watchObservedRunningTime="2025-10-09 10:25:56.302473416 +0000 UTC m=+1242.370655172" Oct 09 10:25:56 crc kubenswrapper[4923]: I1009 10:25:56.322808 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-b2zqk" podStartSLOduration=2.322784593 podStartE2EDuration="2.322784593s" podCreationTimestamp="2025-10-09 10:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:25:56.320112529 +0000 UTC m=+1242.388294285" watchObservedRunningTime="2025-10-09 10:25:56.322784593 +0000 UTC m=+1242.390966349" Oct 09 10:25:57 crc kubenswrapper[4923]: I1009 10:25:57.290512 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerStarted","Data":"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec"} Oct 09 10:25:57 crc kubenswrapper[4923]: I1009 10:25:57.686032 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:25:57 crc kubenswrapper[4923]: I1009 10:25:57.753312 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-9scnz"] Oct 09 10:25:57 crc kubenswrapper[4923]: I1009 10:25:57.758048 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" podUID="6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" containerName="dnsmasq-dns" containerID="cri-o://19ced66d3a8413134b8e05e7b359f29a4b16c6b7ce0d1861cb75903f98fe986b" gracePeriod=10 Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.340805 4923 generic.go:334] "Generic (PLEG): container finished" podID="6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" containerID="19ced66d3a8413134b8e05e7b359f29a4b16c6b7ce0d1861cb75903f98fe986b" exitCode=0 Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.341356 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" event={"ID":"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6","Type":"ContainerDied","Data":"19ced66d3a8413134b8e05e7b359f29a4b16c6b7ce0d1861cb75903f98fe986b"} Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.341431 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" event={"ID":"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6","Type":"ContainerDied","Data":"c2f2f326f7121fbbab6af0b1dc1c82bc8697570485e54b944a8b2d1252a7e62b"} Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.341456 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2f2f326f7121fbbab6af0b1dc1c82bc8697570485e54b944a8b2d1252a7e62b" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.388822 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.565115 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-sb\") pod \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.565606 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-config\") pod \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.565799 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-nb\") pod \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.565914 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-dns-svc\") pod \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.567528 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpbdd\" (UniqueName: \"kubernetes.io/projected/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-kube-api-access-gpbdd\") pod \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\" (UID: \"6ebe8dc7-4f22-41d7-bd64-178ec7a799d6\") " Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.574256 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-kube-api-access-gpbdd" (OuterVolumeSpecName: "kube-api-access-gpbdd") pod "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" (UID: "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6"). InnerVolumeSpecName "kube-api-access-gpbdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.629554 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" (UID: "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.639348 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" (UID: "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.639621 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" (UID: "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.642121 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-config" (OuterVolumeSpecName: "config") pod "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" (UID: "6ebe8dc7-4f22-41d7-bd64-178ec7a799d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.686260 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.687438 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.687463 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.687473 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpbdd\" (UniqueName: \"kubernetes.io/projected/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-kube-api-access-gpbdd\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:58 crc kubenswrapper[4923]: I1009 10:25:58.687484 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.355223 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54974c8ff5-9scnz" Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.355950 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="ceilometer-central-agent" containerID="cri-o://93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660" gracePeriod=30 Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.356114 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="proxy-httpd" containerID="cri-o://71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da" gracePeriod=30 Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.356154 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="sg-core" containerID="cri-o://b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec" gracePeriod=30 Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.356189 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="ceilometer-notification-agent" containerID="cri-o://496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128" gracePeriod=30 Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.357163 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerStarted","Data":"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da"} Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.357698 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.402437 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5633160630000003 podStartE2EDuration="7.402398622s" podCreationTimestamp="2025-10-09 10:25:52 +0000 UTC" firstStartedPulling="2025-10-09 10:25:53.227128695 +0000 UTC m=+1239.295310451" lastFinishedPulling="2025-10-09 10:25:58.066211254 +0000 UTC m=+1244.134393010" observedRunningTime="2025-10-09 10:25:59.385351314 +0000 UTC m=+1245.453533070" watchObservedRunningTime="2025-10-09 10:25:59.402398622 +0000 UTC m=+1245.470580378" Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.422670 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-9scnz"] Oct 09 10:25:59 crc kubenswrapper[4923]: I1009 10:25:59.432074 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54974c8ff5-9scnz"] Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.333111 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421011 4923 generic.go:334] "Generic (PLEG): container finished" podID="38dde525-347c-4d29-bbd7-20a1e7712193" containerID="71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da" exitCode=0 Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421057 4923 generic.go:334] "Generic (PLEG): container finished" podID="38dde525-347c-4d29-bbd7-20a1e7712193" containerID="b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec" exitCode=2 Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421067 4923 generic.go:334] "Generic (PLEG): container finished" podID="38dde525-347c-4d29-bbd7-20a1e7712193" containerID="496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128" exitCode=0 Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421076 4923 generic.go:334] "Generic (PLEG): container finished" podID="38dde525-347c-4d29-bbd7-20a1e7712193" containerID="93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660" exitCode=0 Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421103 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerDied","Data":"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da"} Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421136 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerDied","Data":"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec"} Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421146 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerDied","Data":"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128"} Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421156 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerDied","Data":"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660"} Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421165 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"38dde525-347c-4d29-bbd7-20a1e7712193","Type":"ContainerDied","Data":"7b31e1526307b9273759fa0c24cbee3269a59fa7bdcf529bbb9c646abe668748"} Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421182 4923 scope.go:117] "RemoveContainer" containerID="71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.421381 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.491046 4923 scope.go:117] "RemoveContainer" containerID="b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.529509 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-ceilometer-tls-certs\") pod \"38dde525-347c-4d29-bbd7-20a1e7712193\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.529601 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-config-data\") pod \"38dde525-347c-4d29-bbd7-20a1e7712193\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.529622 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-scripts\") pod \"38dde525-347c-4d29-bbd7-20a1e7712193\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.529662 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-combined-ca-bundle\") pod \"38dde525-347c-4d29-bbd7-20a1e7712193\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.529731 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnwc6\" (UniqueName: \"kubernetes.io/projected/38dde525-347c-4d29-bbd7-20a1e7712193-kube-api-access-dnwc6\") pod \"38dde525-347c-4d29-bbd7-20a1e7712193\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.529921 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-sg-core-conf-yaml\") pod \"38dde525-347c-4d29-bbd7-20a1e7712193\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.529954 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-run-httpd\") pod \"38dde525-347c-4d29-bbd7-20a1e7712193\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.530026 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-log-httpd\") pod \"38dde525-347c-4d29-bbd7-20a1e7712193\" (UID: \"38dde525-347c-4d29-bbd7-20a1e7712193\") " Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.531319 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "38dde525-347c-4d29-bbd7-20a1e7712193" (UID: "38dde525-347c-4d29-bbd7-20a1e7712193"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.533081 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "38dde525-347c-4d29-bbd7-20a1e7712193" (UID: "38dde525-347c-4d29-bbd7-20a1e7712193"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.549148 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-scripts" (OuterVolumeSpecName: "scripts") pod "38dde525-347c-4d29-bbd7-20a1e7712193" (UID: "38dde525-347c-4d29-bbd7-20a1e7712193"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.551242 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38dde525-347c-4d29-bbd7-20a1e7712193-kube-api-access-dnwc6" (OuterVolumeSpecName: "kube-api-access-dnwc6") pod "38dde525-347c-4d29-bbd7-20a1e7712193" (UID: "38dde525-347c-4d29-bbd7-20a1e7712193"). InnerVolumeSpecName "kube-api-access-dnwc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.598939 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "38dde525-347c-4d29-bbd7-20a1e7712193" (UID: "38dde525-347c-4d29-bbd7-20a1e7712193"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.626337 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" path="/var/lib/kubelet/pods/6ebe8dc7-4f22-41d7-bd64-178ec7a799d6/volumes" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.637811 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.638056 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnwc6\" (UniqueName: \"kubernetes.io/projected/38dde525-347c-4d29-bbd7-20a1e7712193-kube-api-access-dnwc6\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.638287 4923 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.638369 4923 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.639634 4923 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/38dde525-347c-4d29-bbd7-20a1e7712193-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.646183 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "38dde525-347c-4d29-bbd7-20a1e7712193" (UID: "38dde525-347c-4d29-bbd7-20a1e7712193"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.682091 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-config-data" (OuterVolumeSpecName: "config-data") pod "38dde525-347c-4d29-bbd7-20a1e7712193" (UID: "38dde525-347c-4d29-bbd7-20a1e7712193"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.708454 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "38dde525-347c-4d29-bbd7-20a1e7712193" (UID: "38dde525-347c-4d29-bbd7-20a1e7712193"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.731550 4923 scope.go:117] "RemoveContainer" containerID="496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.742237 4923 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.742298 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.742311 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38dde525-347c-4d29-bbd7-20a1e7712193-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.760307 4923 scope.go:117] "RemoveContainer" containerID="93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.767575 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.777150 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.790797 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.791256 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="ceilometer-central-agent" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791280 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="ceilometer-central-agent" Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.791302 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" containerName="dnsmasq-dns" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791310 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" containerName="dnsmasq-dns" Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.791326 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="sg-core" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791332 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="sg-core" Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.791346 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="ceilometer-notification-agent" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791352 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="ceilometer-notification-agent" Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.791369 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="proxy-httpd" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791374 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="proxy-httpd" Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.791387 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" containerName="init" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791394 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" containerName="init" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791591 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="ceilometer-central-agent" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791608 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="proxy-httpd" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791617 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="ceilometer-notification-agent" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791629 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" containerName="sg-core" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.791641 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ebe8dc7-4f22-41d7-bd64-178ec7a799d6" containerName="dnsmasq-dns" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.792622 4923 scope.go:117] "RemoveContainer" containerID="71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da" Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.793436 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da\": container with ID starting with 71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da not found: ID does not exist" containerID="71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.793517 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da"} err="failed to get container status \"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da\": rpc error: code = NotFound desc = could not find container \"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da\": container with ID starting with 71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.793563 4923 scope.go:117] "RemoveContainer" containerID="b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec" Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.793979 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec\": container with ID starting with b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec not found: ID does not exist" containerID="b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.794125 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec"} err="failed to get container status \"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec\": rpc error: code = NotFound desc = could not find container \"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec\": container with ID starting with b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.794249 4923 scope.go:117] "RemoveContainer" containerID="496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.794263 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.796496 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128\": container with ID starting with 496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128 not found: ID does not exist" containerID="496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.796538 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128"} err="failed to get container status \"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128\": rpc error: code = NotFound desc = could not find container \"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128\": container with ID starting with 496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128 not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.796562 4923 scope.go:117] "RemoveContainer" containerID="93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660" Oct 09 10:26:00 crc kubenswrapper[4923]: E1009 10:26:00.796955 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660\": container with ID starting with 93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660 not found: ID does not exist" containerID="93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.797011 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660"} err="failed to get container status \"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660\": rpc error: code = NotFound desc = could not find container \"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660\": container with ID starting with 93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660 not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.797052 4923 scope.go:117] "RemoveContainer" containerID="71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.801154 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da"} err="failed to get container status \"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da\": rpc error: code = NotFound desc = could not find container \"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da\": container with ID starting with 71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.801484 4923 scope.go:117] "RemoveContainer" containerID="b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.801204 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.801965 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.802128 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.802844 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec"} err="failed to get container status \"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec\": rpc error: code = NotFound desc = could not find container \"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec\": container with ID starting with b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.802883 4923 scope.go:117] "RemoveContainer" containerID="496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.803323 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128"} err="failed to get container status \"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128\": rpc error: code = NotFound desc = could not find container \"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128\": container with ID starting with 496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128 not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.803449 4923 scope.go:117] "RemoveContainer" containerID="93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.804648 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660"} err="failed to get container status \"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660\": rpc error: code = NotFound desc = could not find container \"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660\": container with ID starting with 93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660 not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.804715 4923 scope.go:117] "RemoveContainer" containerID="71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.806615 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da"} err="failed to get container status \"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da\": rpc error: code = NotFound desc = could not find container \"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da\": container with ID starting with 71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.807195 4923 scope.go:117] "RemoveContainer" containerID="b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.810053 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec"} err="failed to get container status \"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec\": rpc error: code = NotFound desc = could not find container \"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec\": container with ID starting with b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.810109 4923 scope.go:117] "RemoveContainer" containerID="496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.810585 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128"} err="failed to get container status \"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128\": rpc error: code = NotFound desc = could not find container \"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128\": container with ID starting with 496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128 not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.810628 4923 scope.go:117] "RemoveContainer" containerID="93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.812544 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660"} err="failed to get container status \"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660\": rpc error: code = NotFound desc = could not find container \"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660\": container with ID starting with 93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660 not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.812604 4923 scope.go:117] "RemoveContainer" containerID="71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.813176 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da"} err="failed to get container status \"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da\": rpc error: code = NotFound desc = could not find container \"71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da\": container with ID starting with 71af647002255cb1309221a55d6e8f6edc54ba6cee3540b564e0fe6fe2dd90da not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.813207 4923 scope.go:117] "RemoveContainer" containerID="b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.813707 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec"} err="failed to get container status \"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec\": rpc error: code = NotFound desc = could not find container \"b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec\": container with ID starting with b97a47e4cdcc99ebe9d7d8c977cd2151bc4f6adb1288606cdad27416aad503ec not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.813825 4923 scope.go:117] "RemoveContainer" containerID="496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.814151 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128"} err="failed to get container status \"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128\": rpc error: code = NotFound desc = could not find container \"496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128\": container with ID starting with 496731bbc814e51952339866378eebbeb37e43c1f10f0ea72da975e1d7c76128 not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.814235 4923 scope.go:117] "RemoveContainer" containerID="93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.814517 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660"} err="failed to get container status \"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660\": rpc error: code = NotFound desc = could not find container \"93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660\": container with ID starting with 93540e7c30253ba70b2e993ebb75b3c1334134de9a6e2f11fc175930d1340660 not found: ID does not exist" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.824177 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.848127 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.848233 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-run-httpd\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.848259 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-log-httpd\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.848343 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-config-data\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.848379 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvpjs\" (UniqueName: \"kubernetes.io/projected/88b5043d-77b0-40c5-b0d0-c9283f084046-kube-api-access-bvpjs\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.848448 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.848484 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-scripts\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.848512 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950042 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-config-data\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950119 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvpjs\" (UniqueName: \"kubernetes.io/projected/88b5043d-77b0-40c5-b0d0-c9283f084046-kube-api-access-bvpjs\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950194 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950225 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-scripts\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950252 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950279 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950318 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-run-httpd\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950348 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-log-httpd\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950847 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-log-httpd\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.950954 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-run-httpd\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.955429 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.955677 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.966241 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-scripts\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.966281 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.966939 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-config-data\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:00 crc kubenswrapper[4923]: I1009 10:26:00.975204 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvpjs\" (UniqueName: \"kubernetes.io/projected/88b5043d-77b0-40c5-b0d0-c9283f084046-kube-api-access-bvpjs\") pod \"ceilometer-0\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " pod="openstack/ceilometer-0" Oct 09 10:26:01 crc kubenswrapper[4923]: I1009 10:26:01.125294 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 10:26:01 crc kubenswrapper[4923]: W1009 10:26:01.654025 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod88b5043d_77b0_40c5_b0d0_c9283f084046.slice/crio-774aae2b684e8d8c4c268bd7ce85b1a317b7b4a4f0dd3d5b80bf14bc2c3851ca WatchSource:0}: Error finding container 774aae2b684e8d8c4c268bd7ce85b1a317b7b4a4f0dd3d5b80bf14bc2c3851ca: Status 404 returned error can't find the container with id 774aae2b684e8d8c4c268bd7ce85b1a317b7b4a4f0dd3d5b80bf14bc2c3851ca Oct 09 10:26:01 crc kubenswrapper[4923]: I1009 10:26:01.655008 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 10:26:02 crc kubenswrapper[4923]: I1009 10:26:02.462201 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerStarted","Data":"0a26caf24f63cdb362fad2b04c5f6032e42d26a46470aa8536c2d700a687d53d"} Oct 09 10:26:02 crc kubenswrapper[4923]: I1009 10:26:02.462795 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerStarted","Data":"774aae2b684e8d8c4c268bd7ce85b1a317b7b4a4f0dd3d5b80bf14bc2c3851ca"} Oct 09 10:26:02 crc kubenswrapper[4923]: I1009 10:26:02.465134 4923 generic.go:334] "Generic (PLEG): container finished" podID="daf12e8f-fbe9-44dd-aff6-c8033ae56d45" containerID="48f3f03b0b0e33fb4cb6af6c4d9de2cdc3336e631ffa235c713c4136c63a2637" exitCode=0 Oct 09 10:26:02 crc kubenswrapper[4923]: I1009 10:26:02.465199 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b2zqk" event={"ID":"daf12e8f-fbe9-44dd-aff6-c8033ae56d45","Type":"ContainerDied","Data":"48f3f03b0b0e33fb4cb6af6c4d9de2cdc3336e631ffa235c713c4136c63a2637"} Oct 09 10:26:02 crc kubenswrapper[4923]: I1009 10:26:02.614112 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38dde525-347c-4d29-bbd7-20a1e7712193" path="/var/lib/kubelet/pods/38dde525-347c-4d29-bbd7-20a1e7712193/volumes" Oct 09 10:26:03 crc kubenswrapper[4923]: I1009 10:26:03.478453 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerStarted","Data":"96b37b65285ebc87dfd6573dd7e0528c2ad8cb41af2adb2576cefa9387f7bf55"} Oct 09 10:26:03 crc kubenswrapper[4923]: I1009 10:26:03.889542 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.024651 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-combined-ca-bundle\") pod \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.025344 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-config-data\") pod \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.025947 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l66v4\" (UniqueName: \"kubernetes.io/projected/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-kube-api-access-l66v4\") pod \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.025975 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-scripts\") pod \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\" (UID: \"daf12e8f-fbe9-44dd-aff6-c8033ae56d45\") " Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.030181 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-kube-api-access-l66v4" (OuterVolumeSpecName: "kube-api-access-l66v4") pod "daf12e8f-fbe9-44dd-aff6-c8033ae56d45" (UID: "daf12e8f-fbe9-44dd-aff6-c8033ae56d45"). InnerVolumeSpecName "kube-api-access-l66v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.046044 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-scripts" (OuterVolumeSpecName: "scripts") pod "daf12e8f-fbe9-44dd-aff6-c8033ae56d45" (UID: "daf12e8f-fbe9-44dd-aff6-c8033ae56d45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.069006 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "daf12e8f-fbe9-44dd-aff6-c8033ae56d45" (UID: "daf12e8f-fbe9-44dd-aff6-c8033ae56d45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.086103 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-config-data" (OuterVolumeSpecName: "config-data") pod "daf12e8f-fbe9-44dd-aff6-c8033ae56d45" (UID: "daf12e8f-fbe9-44dd-aff6-c8033ae56d45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.128797 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l66v4\" (UniqueName: \"kubernetes.io/projected/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-kube-api-access-l66v4\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.128851 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.128863 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.128872 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/daf12e8f-fbe9-44dd-aff6-c8033ae56d45-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.492154 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerStarted","Data":"599134c26d70dc49eb13fc37f78e169840223afab406f931c906a7ea090e5b92"} Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.494135 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-b2zqk" event={"ID":"daf12e8f-fbe9-44dd-aff6-c8033ae56d45","Type":"ContainerDied","Data":"3194c9bac6ec3175f98ebca7f847b8d826df0946842d3a1c7513e8f14ad6beeb"} Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.494178 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3194c9bac6ec3175f98ebca7f847b8d826df0946842d3a1c7513e8f14ad6beeb" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.494242 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-b2zqk" Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.758990 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.759355 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5eb00329-d879-4dbe-b9ca-25491082df45" containerName="nova-api-log" containerID="cri-o://df7c565a2ba98b9c9a71d71dcec291fecef54730336abc5d1f3be7d4b2eb3fa0" gracePeriod=30 Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.759542 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5eb00329-d879-4dbe-b9ca-25491082df45" containerName="nova-api-api" containerID="cri-o://2d366a97b0924e8d3eeb3478d7fc9dfeec921a6c51bda9d8b02cbb7e579c03a4" gracePeriod=30 Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.768577 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.769229 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="d216cf8d-f6a9-451a-8787-ff3708898dd6" containerName="nova-scheduler-scheduler" containerID="cri-o://6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec" gracePeriod=30 Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.824084 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.824487 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-log" containerID="cri-o://5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080" gracePeriod=30 Oct 09 10:26:04 crc kubenswrapper[4923]: I1009 10:26:04.824922 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-metadata" containerID="cri-o://bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2" gracePeriod=30 Oct 09 10:26:05 crc kubenswrapper[4923]: E1009 10:26:05.334296 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 09 10:26:05 crc kubenswrapper[4923]: E1009 10:26:05.343835 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 09 10:26:05 crc kubenswrapper[4923]: E1009 10:26:05.349801 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 09 10:26:05 crc kubenswrapper[4923]: E1009 10:26:05.349936 4923 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="d216cf8d-f6a9-451a-8787-ff3708898dd6" containerName="nova-scheduler-scheduler" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.510645 4923 generic.go:334] "Generic (PLEG): container finished" podID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerID="5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080" exitCode=143 Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.510745 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40","Type":"ContainerDied","Data":"5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080"} Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.513239 4923 generic.go:334] "Generic (PLEG): container finished" podID="5eb00329-d879-4dbe-b9ca-25491082df45" containerID="2d366a97b0924e8d3eeb3478d7fc9dfeec921a6c51bda9d8b02cbb7e579c03a4" exitCode=0 Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.513269 4923 generic.go:334] "Generic (PLEG): container finished" podID="5eb00329-d879-4dbe-b9ca-25491082df45" containerID="df7c565a2ba98b9c9a71d71dcec291fecef54730336abc5d1f3be7d4b2eb3fa0" exitCode=143 Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.513307 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eb00329-d879-4dbe-b9ca-25491082df45","Type":"ContainerDied","Data":"2d366a97b0924e8d3eeb3478d7fc9dfeec921a6c51bda9d8b02cbb7e579c03a4"} Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.513336 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eb00329-d879-4dbe-b9ca-25491082df45","Type":"ContainerDied","Data":"df7c565a2ba98b9c9a71d71dcec291fecef54730336abc5d1f3be7d4b2eb3fa0"} Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.513345 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5eb00329-d879-4dbe-b9ca-25491082df45","Type":"ContainerDied","Data":"2ff31e1c716d0703ed16cb9d742b3a49620048b700e884d642cb1af13c0687d6"} Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.513355 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ff31e1c716d0703ed16cb9d742b3a49620048b700e884d642cb1af13c0687d6" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.521230 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerStarted","Data":"9befc1b806d4460d9d89e18548c3d0a5a59c230dc9eabfb969a1367cb2ee88ab"} Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.521883 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.523878 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.575110 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.125900048 podStartE2EDuration="5.575085307s" podCreationTimestamp="2025-10-09 10:26:00 +0000 UTC" firstStartedPulling="2025-10-09 10:26:01.658042123 +0000 UTC m=+1247.726223879" lastFinishedPulling="2025-10-09 10:26:05.107227382 +0000 UTC m=+1251.175409138" observedRunningTime="2025-10-09 10:26:05.552557779 +0000 UTC m=+1251.620739545" watchObservedRunningTime="2025-10-09 10:26:05.575085307 +0000 UTC m=+1251.643267063" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.672032 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-config-data\") pod \"5eb00329-d879-4dbe-b9ca-25491082df45\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.672237 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-combined-ca-bundle\") pod \"5eb00329-d879-4dbe-b9ca-25491082df45\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.673251 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-public-tls-certs\") pod \"5eb00329-d879-4dbe-b9ca-25491082df45\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.673405 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eb00329-d879-4dbe-b9ca-25491082df45-logs\") pod \"5eb00329-d879-4dbe-b9ca-25491082df45\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.673835 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5eb00329-d879-4dbe-b9ca-25491082df45-logs" (OuterVolumeSpecName: "logs") pod "5eb00329-d879-4dbe-b9ca-25491082df45" (UID: "5eb00329-d879-4dbe-b9ca-25491082df45"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.674005 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69gs6\" (UniqueName: \"kubernetes.io/projected/5eb00329-d879-4dbe-b9ca-25491082df45-kube-api-access-69gs6\") pod \"5eb00329-d879-4dbe-b9ca-25491082df45\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.674416 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-internal-tls-certs\") pod \"5eb00329-d879-4dbe-b9ca-25491082df45\" (UID: \"5eb00329-d879-4dbe-b9ca-25491082df45\") " Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.675568 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5eb00329-d879-4dbe-b9ca-25491082df45-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.680998 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5eb00329-d879-4dbe-b9ca-25491082df45-kube-api-access-69gs6" (OuterVolumeSpecName: "kube-api-access-69gs6") pod "5eb00329-d879-4dbe-b9ca-25491082df45" (UID: "5eb00329-d879-4dbe-b9ca-25491082df45"). InnerVolumeSpecName "kube-api-access-69gs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.719810 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5eb00329-d879-4dbe-b9ca-25491082df45" (UID: "5eb00329-d879-4dbe-b9ca-25491082df45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.725863 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-config-data" (OuterVolumeSpecName: "config-data") pod "5eb00329-d879-4dbe-b9ca-25491082df45" (UID: "5eb00329-d879-4dbe-b9ca-25491082df45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.744194 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5eb00329-d879-4dbe-b9ca-25491082df45" (UID: "5eb00329-d879-4dbe-b9ca-25491082df45"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.771431 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5eb00329-d879-4dbe-b9ca-25491082df45" (UID: "5eb00329-d879-4dbe-b9ca-25491082df45"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.777226 4923 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.777283 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69gs6\" (UniqueName: \"kubernetes.io/projected/5eb00329-d879-4dbe-b9ca-25491082df45-kube-api-access-69gs6\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.777296 4923 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.777309 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:05 crc kubenswrapper[4923]: I1009 10:26:05.777322 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5eb00329-d879-4dbe-b9ca-25491082df45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.531590 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.583045 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.591608 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.614999 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5eb00329-d879-4dbe-b9ca-25491082df45" path="/var/lib/kubelet/pods/5eb00329-d879-4dbe-b9ca-25491082df45/volumes" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.615817 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 09 10:26:06 crc kubenswrapper[4923]: E1009 10:26:06.616211 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eb00329-d879-4dbe-b9ca-25491082df45" containerName="nova-api-api" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.616227 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eb00329-d879-4dbe-b9ca-25491082df45" containerName="nova-api-api" Oct 09 10:26:06 crc kubenswrapper[4923]: E1009 10:26:06.616258 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf12e8f-fbe9-44dd-aff6-c8033ae56d45" containerName="nova-manage" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.616267 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf12e8f-fbe9-44dd-aff6-c8033ae56d45" containerName="nova-manage" Oct 09 10:26:06 crc kubenswrapper[4923]: E1009 10:26:06.616280 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5eb00329-d879-4dbe-b9ca-25491082df45" containerName="nova-api-log" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.616288 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5eb00329-d879-4dbe-b9ca-25491082df45" containerName="nova-api-log" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.616702 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eb00329-d879-4dbe-b9ca-25491082df45" containerName="nova-api-api" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.616714 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf12e8f-fbe9-44dd-aff6-c8033ae56d45" containerName="nova-manage" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.616726 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5eb00329-d879-4dbe-b9ca-25491082df45" containerName="nova-api-log" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.618229 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.618431 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.632941 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.634091 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.635070 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.805018 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-config-data\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.805078 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.805124 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/555f5900-a091-41e5-982e-6d8e46d1fbe5-logs\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.805150 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.805215 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6k29v\" (UniqueName: \"kubernetes.io/projected/555f5900-a091-41e5-982e-6d8e46d1fbe5-kube-api-access-6k29v\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.805249 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-public-tls-certs\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.907223 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/555f5900-a091-41e5-982e-6d8e46d1fbe5-logs\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.907297 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.907391 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6k29v\" (UniqueName: \"kubernetes.io/projected/555f5900-a091-41e5-982e-6d8e46d1fbe5-kube-api-access-6k29v\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.907433 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-public-tls-certs\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.907512 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-config-data\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.907539 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.907733 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/555f5900-a091-41e5-982e-6d8e46d1fbe5-logs\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.915902 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-public-tls-certs\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.915931 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.915985 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-config-data\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.921478 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/555f5900-a091-41e5-982e-6d8e46d1fbe5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.925889 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6k29v\" (UniqueName: \"kubernetes.io/projected/555f5900-a091-41e5-982e-6d8e46d1fbe5-kube-api-access-6k29v\") pod \"nova-api-0\" (UID: \"555f5900-a091-41e5-982e-6d8e46d1fbe5\") " pod="openstack/nova-api-0" Oct 09 10:26:06 crc kubenswrapper[4923]: I1009 10:26:06.957575 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 10:26:07 crc kubenswrapper[4923]: I1009 10:26:07.438050 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 10:26:07 crc kubenswrapper[4923]: W1009 10:26:07.438480 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod555f5900_a091_41e5_982e_6d8e46d1fbe5.slice/crio-d4f862b44d970ca86303df5a9d6667aca2fde6160128406ddfdf4ea9e88d7a90 WatchSource:0}: Error finding container d4f862b44d970ca86303df5a9d6667aca2fde6160128406ddfdf4ea9e88d7a90: Status 404 returned error can't find the container with id d4f862b44d970ca86303df5a9d6667aca2fde6160128406ddfdf4ea9e88d7a90 Oct 09 10:26:07 crc kubenswrapper[4923]: I1009 10:26:07.544516 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"555f5900-a091-41e5-982e-6d8e46d1fbe5","Type":"ContainerStarted","Data":"d4f862b44d970ca86303df5a9d6667aca2fde6160128406ddfdf4ea9e88d7a90"} Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.319865 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.443510 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-combined-ca-bundle\") pod \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.443580 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rfkr\" (UniqueName: \"kubernetes.io/projected/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-kube-api-access-4rfkr\") pod \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.443759 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-nova-metadata-tls-certs\") pod \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.443857 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-logs\") pod \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.443919 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-config-data\") pod \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\" (UID: \"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40\") " Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.444526 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-logs" (OuterVolumeSpecName: "logs") pod "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" (UID: "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.445165 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-logs\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.450029 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-kube-api-access-4rfkr" (OuterVolumeSpecName: "kube-api-access-4rfkr") pod "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" (UID: "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40"). InnerVolumeSpecName "kube-api-access-4rfkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.478193 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" (UID: "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.480938 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-config-data" (OuterVolumeSpecName: "config-data") pod "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" (UID: "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.514811 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" (UID: "4f7436ef-251e-4b0d-b066-a7f1c2ce7b40"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.547084 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.547133 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.547148 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rfkr\" (UniqueName: \"kubernetes.io/projected/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-kube-api-access-4rfkr\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.547161 4923 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.558185 4923 generic.go:334] "Generic (PLEG): container finished" podID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerID="bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2" exitCode=0 Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.558348 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40","Type":"ContainerDied","Data":"bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2"} Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.558416 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4f7436ef-251e-4b0d-b066-a7f1c2ce7b40","Type":"ContainerDied","Data":"a0446924177b9195a0098ac89936b8a92bdad0a0ac9460f0ac2be3c5e5039b1f"} Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.558475 4923 scope.go:117] "RemoveContainer" containerID="bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.559255 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.563578 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"555f5900-a091-41e5-982e-6d8e46d1fbe5","Type":"ContainerStarted","Data":"ecd85e998f21976051d862959cbe08230cc22ed092d8a44e3f8b8f6aa27d76c5"} Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.563650 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"555f5900-a091-41e5-982e-6d8e46d1fbe5","Type":"ContainerStarted","Data":"52b86a412f41d1fcdaff4c23bc4b13e1857a48dbaa8bf7099bc7d8da1076d01c"} Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.598166 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.598134224 podStartE2EDuration="2.598134224s" podCreationTimestamp="2025-10-09 10:26:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:26:08.58629548 +0000 UTC m=+1254.654477236" watchObservedRunningTime="2025-10-09 10:26:08.598134224 +0000 UTC m=+1254.666315980" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.629719 4923 scope.go:117] "RemoveContainer" containerID="5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.641577 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.648957 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.673899 4923 scope.go:117] "RemoveContainer" containerID="bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2" Oct 09 10:26:08 crc kubenswrapper[4923]: E1009 10:26:08.674916 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2\": container with ID starting with bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2 not found: ID does not exist" containerID="bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.674955 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2"} err="failed to get container status \"bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2\": rpc error: code = NotFound desc = could not find container \"bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2\": container with ID starting with bd6d5684ac77882038c0c8bf81bd29a071fae959eb7fb406c4dba02c1f2a07a2 not found: ID does not exist" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.674983 4923 scope.go:117] "RemoveContainer" containerID="5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080" Oct 09 10:26:08 crc kubenswrapper[4923]: E1009 10:26:08.675433 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080\": container with ID starting with 5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080 not found: ID does not exist" containerID="5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.675453 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080"} err="failed to get container status \"5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080\": rpc error: code = NotFound desc = could not find container \"5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080\": container with ID starting with 5da93350bdb6aaa7abb854a7b74cd58195cdbf53e9232648a65f7267ffb81080 not found: ID does not exist" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.680095 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:26:08 crc kubenswrapper[4923]: E1009 10:26:08.680915 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-metadata" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.680938 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-metadata" Oct 09 10:26:08 crc kubenswrapper[4923]: E1009 10:26:08.680950 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-log" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.680957 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-log" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.681175 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-log" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.681206 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-metadata" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.682588 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.687090 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.687354 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.692113 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.856126 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5e67cba-1bce-4d4a-b1f0-079ebd765787-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.856859 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e67cba-1bce-4d4a-b1f0-079ebd765787-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.856937 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e67cba-1bce-4d4a-b1f0-079ebd765787-config-data\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.857019 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlfsg\" (UniqueName: \"kubernetes.io/projected/e5e67cba-1bce-4d4a-b1f0-079ebd765787-kube-api-access-tlfsg\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.857381 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5e67cba-1bce-4d4a-b1f0-079ebd765787-logs\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.959410 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5e67cba-1bce-4d4a-b1f0-079ebd765787-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.959531 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e67cba-1bce-4d4a-b1f0-079ebd765787-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.959577 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e67cba-1bce-4d4a-b1f0-079ebd765787-config-data\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.959618 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlfsg\" (UniqueName: \"kubernetes.io/projected/e5e67cba-1bce-4d4a-b1f0-079ebd765787-kube-api-access-tlfsg\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.959700 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5e67cba-1bce-4d4a-b1f0-079ebd765787-logs\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.960366 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5e67cba-1bce-4d4a-b1f0-079ebd765787-logs\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.967140 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5e67cba-1bce-4d4a-b1f0-079ebd765787-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.968647 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5e67cba-1bce-4d4a-b1f0-079ebd765787-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.981209 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5e67cba-1bce-4d4a-b1f0-079ebd765787-config-data\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:08 crc kubenswrapper[4923]: I1009 10:26:08.993356 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlfsg\" (UniqueName: \"kubernetes.io/projected/e5e67cba-1bce-4d4a-b1f0-079ebd765787-kube-api-access-tlfsg\") pod \"nova-metadata-0\" (UID: \"e5e67cba-1bce-4d4a-b1f0-079ebd765787\") " pod="openstack/nova-metadata-0" Oct 09 10:26:09 crc kubenswrapper[4923]: I1009 10:26:09.008016 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 10:26:09 crc kubenswrapper[4923]: I1009 10:26:09.522368 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 10:26:09 crc kubenswrapper[4923]: W1009 10:26:09.574114 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5e67cba_1bce_4d4a_b1f0_079ebd765787.slice/crio-7e817206bbe11767b5089633bbfd896c54f05ecbb8462f5b94b177dbeca6a89b WatchSource:0}: Error finding container 7e817206bbe11767b5089633bbfd896c54f05ecbb8462f5b94b177dbeca6a89b: Status 404 returned error can't find the container with id 7e817206bbe11767b5089633bbfd896c54f05ecbb8462f5b94b177dbeca6a89b Oct 09 10:26:09 crc kubenswrapper[4923]: I1009 10:26:09.938550 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.097830 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zrj8\" (UniqueName: \"kubernetes.io/projected/d216cf8d-f6a9-451a-8787-ff3708898dd6-kube-api-access-5zrj8\") pod \"d216cf8d-f6a9-451a-8787-ff3708898dd6\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.098058 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-combined-ca-bundle\") pod \"d216cf8d-f6a9-451a-8787-ff3708898dd6\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.098204 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-config-data\") pod \"d216cf8d-f6a9-451a-8787-ff3708898dd6\" (UID: \"d216cf8d-f6a9-451a-8787-ff3708898dd6\") " Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.102550 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d216cf8d-f6a9-451a-8787-ff3708898dd6-kube-api-access-5zrj8" (OuterVolumeSpecName: "kube-api-access-5zrj8") pod "d216cf8d-f6a9-451a-8787-ff3708898dd6" (UID: "d216cf8d-f6a9-451a-8787-ff3708898dd6"). InnerVolumeSpecName "kube-api-access-5zrj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.130803 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-config-data" (OuterVolumeSpecName: "config-data") pod "d216cf8d-f6a9-451a-8787-ff3708898dd6" (UID: "d216cf8d-f6a9-451a-8787-ff3708898dd6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.134913 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d216cf8d-f6a9-451a-8787-ff3708898dd6" (UID: "d216cf8d-f6a9-451a-8787-ff3708898dd6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.202075 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.202130 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d216cf8d-f6a9-451a-8787-ff3708898dd6-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.202143 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zrj8\" (UniqueName: \"kubernetes.io/projected/d216cf8d-f6a9-451a-8787-ff3708898dd6-kube-api-access-5zrj8\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.595417 4923 generic.go:334] "Generic (PLEG): container finished" podID="d216cf8d-f6a9-451a-8787-ff3708898dd6" containerID="6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec" exitCode=0 Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.595530 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.595515 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d216cf8d-f6a9-451a-8787-ff3708898dd6","Type":"ContainerDied","Data":"6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec"} Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.595696 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d216cf8d-f6a9-451a-8787-ff3708898dd6","Type":"ContainerDied","Data":"1fa52b58814cb507731e9711ffc53469a5da4024577911c6d3a6cbacd3e75bca"} Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.595744 4923 scope.go:117] "RemoveContainer" containerID="6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.625863 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" path="/var/lib/kubelet/pods/4f7436ef-251e-4b0d-b066-a7f1c2ce7b40/volumes" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.626879 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5e67cba-1bce-4d4a-b1f0-079ebd765787","Type":"ContainerStarted","Data":"df8f340894e98a1997c18f96754d7d86627f2a74f6155ad3e838d19d3db65982"} Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.626949 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5e67cba-1bce-4d4a-b1f0-079ebd765787","Type":"ContainerStarted","Data":"5e35cd946872f25db40451b64794afb868d25c52fbb7f2264f5bc4914057f0a5"} Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.626968 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e5e67cba-1bce-4d4a-b1f0-079ebd765787","Type":"ContainerStarted","Data":"7e817206bbe11767b5089633bbfd896c54f05ecbb8462f5b94b177dbeca6a89b"} Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.627900 4923 scope.go:117] "RemoveContainer" containerID="6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec" Oct 09 10:26:10 crc kubenswrapper[4923]: E1009 10:26:10.634053 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec\": container with ID starting with 6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec not found: ID does not exist" containerID="6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.634257 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec"} err="failed to get container status \"6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec\": rpc error: code = NotFound desc = could not find container \"6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec\": container with ID starting with 6c5922bf9f5ce4cce4d531bd07d0261a54da882e221ed22a4acdae1b949e55ec not found: ID does not exist" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.645111 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.645081025 podStartE2EDuration="2.645081025s" podCreationTimestamp="2025-10-09 10:26:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:26:10.634028962 +0000 UTC m=+1256.702210738" watchObservedRunningTime="2025-10-09 10:26:10.645081025 +0000 UTC m=+1256.713262781" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.662247 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.678574 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.690709 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:26:10 crc kubenswrapper[4923]: E1009 10:26:10.691469 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d216cf8d-f6a9-451a-8787-ff3708898dd6" containerName="nova-scheduler-scheduler" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.691502 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d216cf8d-f6a9-451a-8787-ff3708898dd6" containerName="nova-scheduler-scheduler" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.691842 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d216cf8d-f6a9-451a-8787-ff3708898dd6" containerName="nova-scheduler-scheduler" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.692933 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.695866 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.702682 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.716702 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f5090b-66d4-409a-b7d1-013eda9b7f45-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37f5090b-66d4-409a-b7d1-013eda9b7f45\") " pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.716874 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5zq2\" (UniqueName: \"kubernetes.io/projected/37f5090b-66d4-409a-b7d1-013eda9b7f45-kube-api-access-m5zq2\") pod \"nova-scheduler-0\" (UID: \"37f5090b-66d4-409a-b7d1-013eda9b7f45\") " pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.716962 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f5090b-66d4-409a-b7d1-013eda9b7f45-config-data\") pod \"nova-scheduler-0\" (UID: \"37f5090b-66d4-409a-b7d1-013eda9b7f45\") " pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.818581 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f5090b-66d4-409a-b7d1-013eda9b7f45-config-data\") pod \"nova-scheduler-0\" (UID: \"37f5090b-66d4-409a-b7d1-013eda9b7f45\") " pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.818768 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f5090b-66d4-409a-b7d1-013eda9b7f45-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37f5090b-66d4-409a-b7d1-013eda9b7f45\") " pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.818812 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5zq2\" (UniqueName: \"kubernetes.io/projected/37f5090b-66d4-409a-b7d1-013eda9b7f45-kube-api-access-m5zq2\") pod \"nova-scheduler-0\" (UID: \"37f5090b-66d4-409a-b7d1-013eda9b7f45\") " pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.829739 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37f5090b-66d4-409a-b7d1-013eda9b7f45-config-data\") pod \"nova-scheduler-0\" (UID: \"37f5090b-66d4-409a-b7d1-013eda9b7f45\") " pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.829739 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37f5090b-66d4-409a-b7d1-013eda9b7f45-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"37f5090b-66d4-409a-b7d1-013eda9b7f45\") " pod="openstack/nova-scheduler-0" Oct 09 10:26:10 crc kubenswrapper[4923]: I1009 10:26:10.840680 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5zq2\" (UniqueName: \"kubernetes.io/projected/37f5090b-66d4-409a-b7d1-013eda9b7f45-kube-api-access-m5zq2\") pod \"nova-scheduler-0\" (UID: \"37f5090b-66d4-409a-b7d1-013eda9b7f45\") " pod="openstack/nova-scheduler-0" Oct 09 10:26:11 crc kubenswrapper[4923]: I1009 10:26:11.023344 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 10:26:11 crc kubenswrapper[4923]: I1009 10:26:11.548024 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 10:26:11 crc kubenswrapper[4923]: I1009 10:26:11.625472 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37f5090b-66d4-409a-b7d1-013eda9b7f45","Type":"ContainerStarted","Data":"d98b74c862106d1f9565e3cb88a220e58ca79ab0599cc2a459130994526e9c14"} Oct 09 10:26:12 crc kubenswrapper[4923]: I1009 10:26:12.612668 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d216cf8d-f6a9-451a-8787-ff3708898dd6" path="/var/lib/kubelet/pods/d216cf8d-f6a9-451a-8787-ff3708898dd6/volumes" Oct 09 10:26:12 crc kubenswrapper[4923]: I1009 10:26:12.643232 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"37f5090b-66d4-409a-b7d1-013eda9b7f45","Type":"ContainerStarted","Data":"040b100bea2876ed5f544fc44d6fc4f2b047c6959738aa344a05e0b847c83c74"} Oct 09 10:26:12 crc kubenswrapper[4923]: I1009 10:26:12.672143 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.672113291 podStartE2EDuration="2.672113291s" podCreationTimestamp="2025-10-09 10:26:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:26:12.663485123 +0000 UTC m=+1258.731666879" watchObservedRunningTime="2025-10-09 10:26:12.672113291 +0000 UTC m=+1258.740295047" Oct 09 10:26:13 crc kubenswrapper[4923]: I1009 10:26:13.299382 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 09 10:26:13 crc kubenswrapper[4923]: I1009 10:26:13.299492 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="4f7436ef-251e-4b0d-b066-a7f1c2ce7b40" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.183:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 09 10:26:14 crc kubenswrapper[4923]: I1009 10:26:14.008553 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 10:26:14 crc kubenswrapper[4923]: I1009 10:26:14.009829 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 10:26:16 crc kubenswrapper[4923]: I1009 10:26:16.025041 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 09 10:26:16 crc kubenswrapper[4923]: I1009 10:26:16.958273 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 10:26:16 crc kubenswrapper[4923]: I1009 10:26:16.958821 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 10:26:17 crc kubenswrapper[4923]: I1009 10:26:17.975122 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="555f5900-a091-41e5-982e-6d8e46d1fbe5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 10:26:17 crc kubenswrapper[4923]: I1009 10:26:17.975582 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="555f5900-a091-41e5-982e-6d8e46d1fbe5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.193:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 10:26:19 crc kubenswrapper[4923]: I1009 10:26:19.009566 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 09 10:26:19 crc kubenswrapper[4923]: I1009 10:26:19.010017 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 09 10:26:20 crc kubenswrapper[4923]: I1009 10:26:20.027040 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e5e67cba-1bce-4d4a-b1f0-079ebd765787" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 10:26:20 crc kubenswrapper[4923]: I1009 10:26:20.027063 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e5e67cba-1bce-4d4a-b1f0-079ebd765787" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.194:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 10:26:21 crc kubenswrapper[4923]: I1009 10:26:21.024805 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 09 10:26:21 crc kubenswrapper[4923]: I1009 10:26:21.060643 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 09 10:26:21 crc kubenswrapper[4923]: I1009 10:26:21.801467 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 09 10:26:24 crc kubenswrapper[4923]: I1009 10:26:24.599342 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:26:24 crc kubenswrapper[4923]: I1009 10:26:24.599896 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:26:26 crc kubenswrapper[4923]: I1009 10:26:26.968376 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 09 10:26:26 crc kubenswrapper[4923]: I1009 10:26:26.969887 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 09 10:26:26 crc kubenswrapper[4923]: I1009 10:26:26.970482 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 09 10:26:26 crc kubenswrapper[4923]: I1009 10:26:26.970572 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 09 10:26:26 crc kubenswrapper[4923]: I1009 10:26:26.978859 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 09 10:26:26 crc kubenswrapper[4923]: I1009 10:26:26.979323 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 09 10:26:29 crc kubenswrapper[4923]: I1009 10:26:29.022557 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 09 10:26:29 crc kubenswrapper[4923]: I1009 10:26:29.024353 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 09 10:26:29 crc kubenswrapper[4923]: I1009 10:26:29.029202 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 09 10:26:29 crc kubenswrapper[4923]: I1009 10:26:29.842694 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 09 10:26:31 crc kubenswrapper[4923]: I1009 10:26:31.134815 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 09 10:26:41 crc kubenswrapper[4923]: I1009 10:26:41.406114 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 10:26:42 crc kubenswrapper[4923]: I1009 10:26:42.268031 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 10:26:46 crc kubenswrapper[4923]: I1009 10:26:46.330068 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="deeda18d-93f0-47c0-8cdd-358505802a0d" containerName="rabbitmq" containerID="cri-o://238b792fa9e2d3b834704439958b1cd2c20b1c82c8d08987fd20821005b85936" gracePeriod=604796 Oct 09 10:26:46 crc kubenswrapper[4923]: I1009 10:26:46.835720 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="29358129-c9b5-4aca-bd00-0473171b3e96" containerName="rabbitmq" containerID="cri-o://19cfaecaaad009dee38b2f7888556bee9b5a45ec2c3aa7f520e6d1a32e05c87b" gracePeriod=604796 Oct 09 10:26:48 crc kubenswrapper[4923]: I1009 10:26:48.605106 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="deeda18d-93f0-47c0-8cdd-358505802a0d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 09 10:26:49 crc kubenswrapper[4923]: I1009 10:26:49.003881 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="29358129-c9b5-4aca-bd00-0473171b3e96" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.101:5671: connect: connection refused" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.067289 4923 generic.go:334] "Generic (PLEG): container finished" podID="29358129-c9b5-4aca-bd00-0473171b3e96" containerID="19cfaecaaad009dee38b2f7888556bee9b5a45ec2c3aa7f520e6d1a32e05c87b" exitCode=0 Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.068449 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29358129-c9b5-4aca-bd00-0473171b3e96","Type":"ContainerDied","Data":"19cfaecaaad009dee38b2f7888556bee9b5a45ec2c3aa7f520e6d1a32e05c87b"} Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.074406 4923 generic.go:334] "Generic (PLEG): container finished" podID="deeda18d-93f0-47c0-8cdd-358505802a0d" containerID="238b792fa9e2d3b834704439958b1cd2c20b1c82c8d08987fd20821005b85936" exitCode=0 Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.074435 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"deeda18d-93f0-47c0-8cdd-358505802a0d","Type":"ContainerDied","Data":"238b792fa9e2d3b834704439958b1cd2c20b1c82c8d08987fd20821005b85936"} Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.074454 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"deeda18d-93f0-47c0-8cdd-358505802a0d","Type":"ContainerDied","Data":"af175916751b1f0164efb54579af39beb3b4e2d64eca40acb107ac4195578305"} Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.074467 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="af175916751b1f0164efb54579af39beb3b4e2d64eca40acb107ac4195578305" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.096176 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.242751 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8kw9\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-kube-api-access-h8kw9\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.242904 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/deeda18d-93f0-47c0-8cdd-358505802a0d-pod-info\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.242947 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/deeda18d-93f0-47c0-8cdd-358505802a0d-erlang-cookie-secret\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.243005 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-tls\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.243063 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-server-conf\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.243094 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-config-data\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.243123 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.243171 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-plugins\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.243215 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-confd\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.243406 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-erlang-cookie\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.243458 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-plugins-conf\") pod \"deeda18d-93f0-47c0-8cdd-358505802a0d\" (UID: \"deeda18d-93f0-47c0-8cdd-358505802a0d\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.245066 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.250220 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.255135 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.263071 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.266967 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/deeda18d-93f0-47c0-8cdd-358505802a0d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.267035 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-kube-api-access-h8kw9" (OuterVolumeSpecName: "kube-api-access-h8kw9") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "kube-api-access-h8kw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.267067 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.269499 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/deeda18d-93f0-47c0-8cdd-358505802a0d-pod-info" (OuterVolumeSpecName: "pod-info") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.314128 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-config-data" (OuterVolumeSpecName: "config-data") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.345553 4923 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/deeda18d-93f0-47c0-8cdd-358505802a0d-pod-info\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.346103 4923 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/deeda18d-93f0-47c0-8cdd-358505802a0d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.346199 4923 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.346280 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.346401 4923 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.346508 4923 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.346596 4923 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.346684 4923 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.346783 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8kw9\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-kube-api-access-h8kw9\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.375605 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-server-conf" (OuterVolumeSpecName: "server-conf") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.376075 4923 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.448254 4923 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/deeda18d-93f0-47c0-8cdd-358505802a0d-server-conf\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.448309 4923 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.455789 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "deeda18d-93f0-47c0-8cdd-358505802a0d" (UID: "deeda18d-93f0-47c0-8cdd-358505802a0d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.485204 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.559271 4923 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/deeda18d-93f0-47c0-8cdd-358505802a0d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.660757 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29358129-c9b5-4aca-bd00-0473171b3e96-pod-info\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.660917 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-plugins-conf\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.661021 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-plugins\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.661072 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.661709 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29358129-c9b5-4aca-bd00-0473171b3e96-erlang-cookie-secret\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.662169 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-config-data\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.662213 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-server-conf\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.662240 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgv9h\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-kube-api-access-cgv9h\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.662258 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-tls\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.662277 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-erlang-cookie\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.662304 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-confd\") pod \"29358129-c9b5-4aca-bd00-0473171b3e96\" (UID: \"29358129-c9b5-4aca-bd00-0473171b3e96\") " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.661726 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.661904 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.666107 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29358129-c9b5-4aca-bd00-0473171b3e96-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.667030 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.667907 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "persistence") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.668600 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.670898 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/29358129-c9b5-4aca-bd00-0473171b3e96-pod-info" (OuterVolumeSpecName: "pod-info") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.672275 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-kube-api-access-cgv9h" (OuterVolumeSpecName: "kube-api-access-cgv9h") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "kube-api-access-cgv9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.696395 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-config-data" (OuterVolumeSpecName: "config-data") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.717822 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-server-conf" (OuterVolumeSpecName: "server-conf") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770628 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgv9h\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-kube-api-access-cgv9h\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770714 4923 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770732 4923 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770745 4923 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/29358129-c9b5-4aca-bd00-0473171b3e96-pod-info\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770779 4923 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770792 4923 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770837 4923 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770851 4923 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/29358129-c9b5-4aca-bd00-0473171b3e96-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770865 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.770880 4923 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/29358129-c9b5-4aca-bd00-0473171b3e96-server-conf\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.793934 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "29358129-c9b5-4aca-bd00-0473171b3e96" (UID: "29358129-c9b5-4aca-bd00-0473171b3e96"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.796843 4923 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.872327 4923 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/29358129-c9b5-4aca-bd00-0473171b3e96-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:53 crc kubenswrapper[4923]: I1009 10:26:53.872356 4923 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.087988 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.087988 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.088018 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"29358129-c9b5-4aca-bd00-0473171b3e96","Type":"ContainerDied","Data":"294c65561254a0bddd82c5a3b05d485133d49c5ae0f5468cc593fa6993694739"} Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.088929 4923 scope.go:117] "RemoveContainer" containerID="19cfaecaaad009dee38b2f7888556bee9b5a45ec2c3aa7f520e6d1a32e05c87b" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.130131 4923 scope.go:117] "RemoveContainer" containerID="863db0b0cc594cea2e1a5f5b6284f95afd669dcec8c8ece965c04e9aa52c6f22" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.152437 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.173915 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.186573 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.201655 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.256462 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 10:26:54 crc kubenswrapper[4923]: E1009 10:26:54.258359 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29358129-c9b5-4aca-bd00-0473171b3e96" containerName="setup-container" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.258401 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="29358129-c9b5-4aca-bd00-0473171b3e96" containerName="setup-container" Oct 09 10:26:54 crc kubenswrapper[4923]: E1009 10:26:54.258411 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeda18d-93f0-47c0-8cdd-358505802a0d" containerName="rabbitmq" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.258417 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeda18d-93f0-47c0-8cdd-358505802a0d" containerName="rabbitmq" Oct 09 10:26:54 crc kubenswrapper[4923]: E1009 10:26:54.258454 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29358129-c9b5-4aca-bd00-0473171b3e96" containerName="rabbitmq" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.258461 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="29358129-c9b5-4aca-bd00-0473171b3e96" containerName="rabbitmq" Oct 09 10:26:54 crc kubenswrapper[4923]: E1009 10:26:54.258502 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="deeda18d-93f0-47c0-8cdd-358505802a0d" containerName="setup-container" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.258511 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="deeda18d-93f0-47c0-8cdd-358505802a0d" containerName="setup-container" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.258883 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="29358129-c9b5-4aca-bd00-0473171b3e96" containerName="rabbitmq" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.258929 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="deeda18d-93f0-47c0-8cdd-358505802a0d" containerName="rabbitmq" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.261695 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.281328 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.281627 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.281818 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.282114 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-g4rg7" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.282268 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.282604 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.282606 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.299302 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.307943 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.312809 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.318982 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.319065 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.319290 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.319376 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.319496 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-frbbl" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.319567 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.319697 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.329914 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384032 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b1b005e0-1c67-42a3-97aa-3222f4218dca-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384121 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384193 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7z6g\" (UniqueName: \"kubernetes.io/projected/b1b005e0-1c67-42a3-97aa-3222f4218dca-kube-api-access-g7z6g\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384266 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384297 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384333 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b1b005e0-1c67-42a3-97aa-3222f4218dca-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384366 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384410 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b1b005e0-1c67-42a3-97aa-3222f4218dca-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384483 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384511 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b1b005e0-1c67-42a3-97aa-3222f4218dca-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.384555 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1b005e0-1c67-42a3-97aa-3222f4218dca-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.486920 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8dbce0ef-9426-4065-9749-633e7a5e4248-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.486987 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487008 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8dbce0ef-9426-4065-9749-633e7a5e4248-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487037 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487125 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487156 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b1b005e0-1c67-42a3-97aa-3222f4218dca-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487182 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8dbce0ef-9426-4065-9749-633e7a5e4248-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487206 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8dbce0ef-9426-4065-9749-633e7a5e4248-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487231 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtgpx\" (UniqueName: \"kubernetes.io/projected/8dbce0ef-9426-4065-9749-633e7a5e4248-kube-api-access-gtgpx\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487250 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1b005e0-1c67-42a3-97aa-3222f4218dca-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487295 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b1b005e0-1c67-42a3-97aa-3222f4218dca-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487322 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487343 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7z6g\" (UniqueName: \"kubernetes.io/projected/b1b005e0-1c67-42a3-97aa-3222f4218dca-kube-api-access-g7z6g\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487368 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487391 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487409 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487432 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b1b005e0-1c67-42a3-97aa-3222f4218dca-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487459 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487486 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487503 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8dbce0ef-9426-4065-9749-633e7a5e4248-config-data\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487522 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b1b005e0-1c67-42a3-97aa-3222f4218dca-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.487539 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.488502 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b1b005e0-1c67-42a3-97aa-3222f4218dca-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.488568 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.488752 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.489509 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b1b005e0-1c67-42a3-97aa-3222f4218dca-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.489814 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.491121 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b1b005e0-1c67-42a3-97aa-3222f4218dca-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.494952 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b1b005e0-1c67-42a3-97aa-3222f4218dca-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.495053 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.495293 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b1b005e0-1c67-42a3-97aa-3222f4218dca-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.495977 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b1b005e0-1c67-42a3-97aa-3222f4218dca-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.512075 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7z6g\" (UniqueName: \"kubernetes.io/projected/b1b005e0-1c67-42a3-97aa-3222f4218dca-kube-api-access-g7z6g\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.516309 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"b1b005e0-1c67-42a3-97aa-3222f4218dca\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.589394 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8dbce0ef-9426-4065-9749-633e7a5e4248-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590301 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8dbce0ef-9426-4065-9749-633e7a5e4248-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590333 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtgpx\" (UniqueName: \"kubernetes.io/projected/8dbce0ef-9426-4065-9749-633e7a5e4248-kube-api-access-gtgpx\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590421 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590469 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590492 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8dbce0ef-9426-4065-9749-633e7a5e4248-config-data\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590561 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590587 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8dbce0ef-9426-4065-9749-633e7a5e4248-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590615 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590631 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8dbce0ef-9426-4065-9749-633e7a5e4248-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.590653 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.591061 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.591315 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.592389 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8dbce0ef-9426-4065-9749-633e7a5e4248-config-data\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.592561 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/8dbce0ef-9426-4065-9749-633e7a5e4248-server-conf\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.592684 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.593195 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/8dbce0ef-9426-4065-9749-633e7a5e4248-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.595996 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/8dbce0ef-9426-4065-9749-633e7a5e4248-pod-info\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.597094 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.597432 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/8dbce0ef-9426-4065-9749-633e7a5e4248-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.610466 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.612949 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/8dbce0ef-9426-4065-9749-633e7a5e4248-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.613333 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtgpx\" (UniqueName: \"kubernetes.io/projected/8dbce0ef-9426-4065-9749-633e7a5e4248-kube-api-access-gtgpx\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.623416 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.623479 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.625529 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29358129-c9b5-4aca-bd00-0473171b3e96" path="/var/lib/kubelet/pods/29358129-c9b5-4aca-bd00-0473171b3e96/volumes" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.626439 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="deeda18d-93f0-47c0-8cdd-358505802a0d" path="/var/lib/kubelet/pods/deeda18d-93f0-47c0-8cdd-358505802a0d/volumes" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.644995 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"8dbce0ef-9426-4065-9749-633e7a5e4248\") " pod="openstack/rabbitmq-server-0" Oct 09 10:26:54 crc kubenswrapper[4923]: I1009 10:26:54.959032 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 10:26:55 crc kubenswrapper[4923]: I1009 10:26:55.266676 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 10:26:55 crc kubenswrapper[4923]: I1009 10:26:55.641960 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 10:26:56 crc kubenswrapper[4923]: I1009 10:26:56.163350 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b1b005e0-1c67-42a3-97aa-3222f4218dca","Type":"ContainerStarted","Data":"75c2515c3ca7a3202859a442fd9d146c0b0d18979e6549c267751def1bacfde1"} Oct 09 10:26:56 crc kubenswrapper[4923]: I1009 10:26:56.166823 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dbce0ef-9426-4065-9749-633e7a5e4248","Type":"ContainerStarted","Data":"45cdec522fe3096074ec4ae0e5a595dc5a17307dc49c6fb8b5a02c4fa0c15608"} Oct 09 10:26:57 crc kubenswrapper[4923]: I1009 10:26:57.216049 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b1b005e0-1c67-42a3-97aa-3222f4218dca","Type":"ContainerStarted","Data":"0109c70360c5b74727f7709bfb9a3d96859ab0153413ac79c881c0c607274c5f"} Oct 09 10:26:58 crc kubenswrapper[4923]: I1009 10:26:58.231351 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dbce0ef-9426-4065-9749-633e7a5e4248","Type":"ContainerStarted","Data":"02235ef01e28456a3525b1efd2f07ad05b46fb1d8584d2a2f853eebfec825646"} Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.787627 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-d9d4l"] Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.790221 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.793844 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.801773 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-d9d4l"] Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.851425 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-config\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.851521 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-openstack-edpm-ipam\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.851555 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj8km\" (UniqueName: \"kubernetes.io/projected/cf99ed5d-244e-4006-ac58-df4cc63ffa11-kube-api-access-fj8km\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.851598 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-dns-svc\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.851660 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-sb\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.851711 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-nb\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.953980 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-sb\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.954089 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-nb\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.954188 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-config\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.954240 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-openstack-edpm-ipam\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.954270 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj8km\" (UniqueName: \"kubernetes.io/projected/cf99ed5d-244e-4006-ac58-df4cc63ffa11-kube-api-access-fj8km\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.954305 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-dns-svc\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.955191 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-dns-svc\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.955235 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-sb\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.955337 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-config\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.955964 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-openstack-edpm-ipam\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.956132 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-nb\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:00 crc kubenswrapper[4923]: I1009 10:27:00.978275 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj8km\" (UniqueName: \"kubernetes.io/projected/cf99ed5d-244e-4006-ac58-df4cc63ffa11-kube-api-access-fj8km\") pod \"dnsmasq-dns-5745cbd8d7-d9d4l\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:01 crc kubenswrapper[4923]: I1009 10:27:01.130486 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:01 crc kubenswrapper[4923]: I1009 10:27:01.746004 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-d9d4l"] Oct 09 10:27:02 crc kubenswrapper[4923]: I1009 10:27:02.278703 4923 generic.go:334] "Generic (PLEG): container finished" podID="cf99ed5d-244e-4006-ac58-df4cc63ffa11" containerID="feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24" exitCode=0 Oct 09 10:27:02 crc kubenswrapper[4923]: I1009 10:27:02.278806 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" event={"ID":"cf99ed5d-244e-4006-ac58-df4cc63ffa11","Type":"ContainerDied","Data":"feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24"} Oct 09 10:27:02 crc kubenswrapper[4923]: I1009 10:27:02.278900 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" event={"ID":"cf99ed5d-244e-4006-ac58-df4cc63ffa11","Type":"ContainerStarted","Data":"91805dd7e56fb6b62c66f3c450216e782227301d1d7d4aca7e21c54069fd0067"} Oct 09 10:27:03 crc kubenswrapper[4923]: I1009 10:27:03.289134 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" event={"ID":"cf99ed5d-244e-4006-ac58-df4cc63ffa11","Type":"ContainerStarted","Data":"8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561"} Oct 09 10:27:03 crc kubenswrapper[4923]: I1009 10:27:03.291732 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:03 crc kubenswrapper[4923]: I1009 10:27:03.315054 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" podStartSLOduration=3.31502992 podStartE2EDuration="3.31502992s" podCreationTimestamp="2025-10-09 10:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:27:03.311077832 +0000 UTC m=+1309.379259588" watchObservedRunningTime="2025-10-09 10:27:03.31502992 +0000 UTC m=+1309.383211676" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.133010 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.200246 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869677f947-lv65r"] Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.200565 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-869677f947-lv65r" podUID="462961cd-6180-4e0a-84b6-cae4c483c869" containerName="dnsmasq-dns" containerID="cri-o://472feb9272ddee5f54959bd6693c5376a5402b3f2a8b8d70965328c54d68b419" gracePeriod=10 Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.390133 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f5d87575-vkntz"] Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.394028 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.401153 4923 generic.go:334] "Generic (PLEG): container finished" podID="462961cd-6180-4e0a-84b6-cae4c483c869" containerID="472feb9272ddee5f54959bd6693c5376a5402b3f2a8b8d70965328c54d68b419" exitCode=0 Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.401224 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-lv65r" event={"ID":"462961cd-6180-4e0a-84b6-cae4c483c869","Type":"ContainerDied","Data":"472feb9272ddee5f54959bd6693c5376a5402b3f2a8b8d70965328c54d68b419"} Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.419389 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f5d87575-vkntz"] Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.517433 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc5v9\" (UniqueName: \"kubernetes.io/projected/1fd11e59-1ac0-4158-b2f9-2286825f3979-kube-api-access-pc5v9\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.517527 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-config\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.517773 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.517855 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-openstack-edpm-ipam\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.517951 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-dns-svc\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.518093 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.621405 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc5v9\" (UniqueName: \"kubernetes.io/projected/1fd11e59-1ac0-4158-b2f9-2286825f3979-kube-api-access-pc5v9\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.621841 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-config\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.621879 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.621899 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-openstack-edpm-ipam\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.621933 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-dns-svc\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.621965 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.622910 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-config\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.623499 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-sb\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.623684 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-dns-svc\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.623730 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-nb\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.624175 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-openstack-edpm-ipam\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.646611 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc5v9\" (UniqueName: \"kubernetes.io/projected/1fd11e59-1ac0-4158-b2f9-2286825f3979-kube-api-access-pc5v9\") pod \"dnsmasq-dns-5f5d87575-vkntz\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.733237 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:11 crc kubenswrapper[4923]: I1009 10:27:11.887996 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.029991 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-dns-svc\") pod \"462961cd-6180-4e0a-84b6-cae4c483c869\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.030145 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-sb\") pod \"462961cd-6180-4e0a-84b6-cae4c483c869\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.031426 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7z7sv\" (UniqueName: \"kubernetes.io/projected/462961cd-6180-4e0a-84b6-cae4c483c869-kube-api-access-7z7sv\") pod \"462961cd-6180-4e0a-84b6-cae4c483c869\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.031474 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-config\") pod \"462961cd-6180-4e0a-84b6-cae4c483c869\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.031506 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-nb\") pod \"462961cd-6180-4e0a-84b6-cae4c483c869\" (UID: \"462961cd-6180-4e0a-84b6-cae4c483c869\") " Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.040376 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/462961cd-6180-4e0a-84b6-cae4c483c869-kube-api-access-7z7sv" (OuterVolumeSpecName: "kube-api-access-7z7sv") pod "462961cd-6180-4e0a-84b6-cae4c483c869" (UID: "462961cd-6180-4e0a-84b6-cae4c483c869"). InnerVolumeSpecName "kube-api-access-7z7sv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.089683 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-config" (OuterVolumeSpecName: "config") pod "462961cd-6180-4e0a-84b6-cae4c483c869" (UID: "462961cd-6180-4e0a-84b6-cae4c483c869"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.094299 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "462961cd-6180-4e0a-84b6-cae4c483c869" (UID: "462961cd-6180-4e0a-84b6-cae4c483c869"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.095091 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "462961cd-6180-4e0a-84b6-cae4c483c869" (UID: "462961cd-6180-4e0a-84b6-cae4c483c869"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.105228 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "462961cd-6180-4e0a-84b6-cae4c483c869" (UID: "462961cd-6180-4e0a-84b6-cae4c483c869"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.134570 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7z7sv\" (UniqueName: \"kubernetes.io/projected/462961cd-6180-4e0a-84b6-cae4c483c869-kube-api-access-7z7sv\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.134627 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.134645 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.134661 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.134673 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/462961cd-6180-4e0a-84b6-cae4c483c869-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.262973 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f5d87575-vkntz"] Oct 09 10:27:12 crc kubenswrapper[4923]: W1009 10:27:12.268707 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fd11e59_1ac0_4158_b2f9_2286825f3979.slice/crio-cac03d949bcbffb8cb9ad15776309518356bed3840e25f4df126d5fcc958b1f9 WatchSource:0}: Error finding container cac03d949bcbffb8cb9ad15776309518356bed3840e25f4df126d5fcc958b1f9: Status 404 returned error can't find the container with id cac03d949bcbffb8cb9ad15776309518356bed3840e25f4df126d5fcc958b1f9 Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.419365 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869677f947-lv65r" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.419361 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869677f947-lv65r" event={"ID":"462961cd-6180-4e0a-84b6-cae4c483c869","Type":"ContainerDied","Data":"aba00e6aa229b8364cea92cd6ba987eb92800d6ee5be5d3809f631ac2278b1b0"} Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.419488 4923 scope.go:117] "RemoveContainer" containerID="472feb9272ddee5f54959bd6693c5376a5402b3f2a8b8d70965328c54d68b419" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.421582 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" event={"ID":"1fd11e59-1ac0-4158-b2f9-2286825f3979","Type":"ContainerStarted","Data":"cac03d949bcbffb8cb9ad15776309518356bed3840e25f4df126d5fcc958b1f9"} Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.465598 4923 scope.go:117] "RemoveContainer" containerID="10448a2d329eefb5a66ecf2d245104d0fe16f581201818ca1ddc07314ba7554c" Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.507533 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869677f947-lv65r"] Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.516682 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869677f947-lv65r"] Oct 09 10:27:12 crc kubenswrapper[4923]: I1009 10:27:12.614439 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="462961cd-6180-4e0a-84b6-cae4c483c869" path="/var/lib/kubelet/pods/462961cd-6180-4e0a-84b6-cae4c483c869/volumes" Oct 09 10:27:13 crc kubenswrapper[4923]: I1009 10:27:13.435518 4923 generic.go:334] "Generic (PLEG): container finished" podID="1fd11e59-1ac0-4158-b2f9-2286825f3979" containerID="4d68ac044f82edb0a9fe5ff23743b1a235111e05923378e0215b34c12c1d6fdc" exitCode=0 Oct 09 10:27:13 crc kubenswrapper[4923]: I1009 10:27:13.435626 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" event={"ID":"1fd11e59-1ac0-4158-b2f9-2286825f3979","Type":"ContainerDied","Data":"4d68ac044f82edb0a9fe5ff23743b1a235111e05923378e0215b34c12c1d6fdc"} Oct 09 10:27:14 crc kubenswrapper[4923]: I1009 10:27:14.448641 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" event={"ID":"1fd11e59-1ac0-4158-b2f9-2286825f3979","Type":"ContainerStarted","Data":"48d3c93b11239e1079e647ef75ae7350edd86932e1c1e41d20c986dc67050ccb"} Oct 09 10:27:14 crc kubenswrapper[4923]: I1009 10:27:14.480397 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" podStartSLOduration=3.480373474 podStartE2EDuration="3.480373474s" podCreationTimestamp="2025-10-09 10:27:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:27:14.47255466 +0000 UTC m=+1320.540736426" watchObservedRunningTime="2025-10-09 10:27:14.480373474 +0000 UTC m=+1320.548555220" Oct 09 10:27:15 crc kubenswrapper[4923]: I1009 10:27:15.460399 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:16 crc kubenswrapper[4923]: I1009 10:27:16.494386 4923 scope.go:117] "RemoveContainer" containerID="2049d58a742c0f3521d999b5b6d1678f1c3179e14b2ad167f60ebabd76e38a1f" Oct 09 10:27:21 crc kubenswrapper[4923]: I1009 10:27:21.734908 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 10:27:21 crc kubenswrapper[4923]: I1009 10:27:21.805000 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-d9d4l"] Oct 09 10:27:21 crc kubenswrapper[4923]: I1009 10:27:21.805354 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" podUID="cf99ed5d-244e-4006-ac58-df4cc63ffa11" containerName="dnsmasq-dns" containerID="cri-o://8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561" gracePeriod=10 Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.384199 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.509164 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-dns-svc\") pod \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.509324 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fj8km\" (UniqueName: \"kubernetes.io/projected/cf99ed5d-244e-4006-ac58-df4cc63ffa11-kube-api-access-fj8km\") pod \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.509379 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-openstack-edpm-ipam\") pod \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.509440 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-sb\") pod \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.509498 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-nb\") pod \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.509523 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-config\") pod \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\" (UID: \"cf99ed5d-244e-4006-ac58-df4cc63ffa11\") " Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.516577 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf99ed5d-244e-4006-ac58-df4cc63ffa11-kube-api-access-fj8km" (OuterVolumeSpecName: "kube-api-access-fj8km") pod "cf99ed5d-244e-4006-ac58-df4cc63ffa11" (UID: "cf99ed5d-244e-4006-ac58-df4cc63ffa11"). InnerVolumeSpecName "kube-api-access-fj8km". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.537868 4923 generic.go:334] "Generic (PLEG): container finished" podID="cf99ed5d-244e-4006-ac58-df4cc63ffa11" containerID="8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561" exitCode=0 Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.537936 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" event={"ID":"cf99ed5d-244e-4006-ac58-df4cc63ffa11","Type":"ContainerDied","Data":"8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561"} Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.537981 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" event={"ID":"cf99ed5d-244e-4006-ac58-df4cc63ffa11","Type":"ContainerDied","Data":"91805dd7e56fb6b62c66f3c450216e782227301d1d7d4aca7e21c54069fd0067"} Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.538004 4923 scope.go:117] "RemoveContainer" containerID="8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.538402 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5745cbd8d7-d9d4l" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.572772 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cf99ed5d-244e-4006-ac58-df4cc63ffa11" (UID: "cf99ed5d-244e-4006-ac58-df4cc63ffa11"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.574016 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "cf99ed5d-244e-4006-ac58-df4cc63ffa11" (UID: "cf99ed5d-244e-4006-ac58-df4cc63ffa11"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.581144 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cf99ed5d-244e-4006-ac58-df4cc63ffa11" (UID: "cf99ed5d-244e-4006-ac58-df4cc63ffa11"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.581241 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-config" (OuterVolumeSpecName: "config") pod "cf99ed5d-244e-4006-ac58-df4cc63ffa11" (UID: "cf99ed5d-244e-4006-ac58-df4cc63ffa11"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.599672 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cf99ed5d-244e-4006-ac58-df4cc63ffa11" (UID: "cf99ed5d-244e-4006-ac58-df4cc63ffa11"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.612261 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.612313 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fj8km\" (UniqueName: \"kubernetes.io/projected/cf99ed5d-244e-4006-ac58-df4cc63ffa11-kube-api-access-fj8km\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.612332 4923 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.612344 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.612482 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.612515 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cf99ed5d-244e-4006-ac58-df4cc63ffa11-config\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.671648 4923 scope.go:117] "RemoveContainer" containerID="feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.702581 4923 scope.go:117] "RemoveContainer" containerID="8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561" Oct 09 10:27:22 crc kubenswrapper[4923]: E1009 10:27:22.703090 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561\": container with ID starting with 8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561 not found: ID does not exist" containerID="8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.703180 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561"} err="failed to get container status \"8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561\": rpc error: code = NotFound desc = could not find container \"8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561\": container with ID starting with 8eb6d2bd44efecc3ac29a50ab77c345a0d93219cfcc4fdd1e89889a1d1148561 not found: ID does not exist" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.703216 4923 scope.go:117] "RemoveContainer" containerID="feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24" Oct 09 10:27:22 crc kubenswrapper[4923]: E1009 10:27:22.703821 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24\": container with ID starting with feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24 not found: ID does not exist" containerID="feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.703880 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24"} err="failed to get container status \"feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24\": rpc error: code = NotFound desc = could not find container \"feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24\": container with ID starting with feb393585dd65af55eb0cc0f4b07830f388b9a0198352c9956a7c2b9aa532c24 not found: ID does not exist" Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.871333 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-d9d4l"] Oct 09 10:27:22 crc kubenswrapper[4923]: I1009 10:27:22.883134 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5745cbd8d7-d9d4l"] Oct 09 10:27:24 crc kubenswrapper[4923]: I1009 10:27:24.599404 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:27:24 crc kubenswrapper[4923]: I1009 10:27:24.600963 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:27:24 crc kubenswrapper[4923]: I1009 10:27:24.614472 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf99ed5d-244e-4006-ac58-df4cc63ffa11" path="/var/lib/kubelet/pods/cf99ed5d-244e-4006-ac58-df4cc63ffa11/volumes" Oct 09 10:27:24 crc kubenswrapper[4923]: I1009 10:27:24.615474 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:27:24 crc kubenswrapper[4923]: I1009 10:27:24.616224 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb8062901f893bbba9ad244080627bdc62960c7f3035d2ee9483ebb1559dcc72"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:27:24 crc kubenswrapper[4923]: I1009 10:27:24.616309 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://cb8062901f893bbba9ad244080627bdc62960c7f3035d2ee9483ebb1559dcc72" gracePeriod=600 Oct 09 10:27:25 crc kubenswrapper[4923]: I1009 10:27:25.576713 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="cb8062901f893bbba9ad244080627bdc62960c7f3035d2ee9483ebb1559dcc72" exitCode=0 Oct 09 10:27:25 crc kubenswrapper[4923]: I1009 10:27:25.576796 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"cb8062901f893bbba9ad244080627bdc62960c7f3035d2ee9483ebb1559dcc72"} Oct 09 10:27:25 crc kubenswrapper[4923]: I1009 10:27:25.577291 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487"} Oct 09 10:27:25 crc kubenswrapper[4923]: I1009 10:27:25.577323 4923 scope.go:117] "RemoveContainer" containerID="ff1594437e0fe053768cc52266cdd8b6d71b1a9aae0c236740cf019156ef7f8f" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.657419 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5"] Oct 09 10:27:27 crc kubenswrapper[4923]: E1009 10:27:27.658936 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462961cd-6180-4e0a-84b6-cae4c483c869" containerName="init" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.658957 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="462961cd-6180-4e0a-84b6-cae4c483c869" containerName="init" Oct 09 10:27:27 crc kubenswrapper[4923]: E1009 10:27:27.658985 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462961cd-6180-4e0a-84b6-cae4c483c869" containerName="dnsmasq-dns" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.658994 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="462961cd-6180-4e0a-84b6-cae4c483c869" containerName="dnsmasq-dns" Oct 09 10:27:27 crc kubenswrapper[4923]: E1009 10:27:27.659020 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf99ed5d-244e-4006-ac58-df4cc63ffa11" containerName="dnsmasq-dns" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.659029 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf99ed5d-244e-4006-ac58-df4cc63ffa11" containerName="dnsmasq-dns" Oct 09 10:27:27 crc kubenswrapper[4923]: E1009 10:27:27.659062 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf99ed5d-244e-4006-ac58-df4cc63ffa11" containerName="init" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.659071 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf99ed5d-244e-4006-ac58-df4cc63ffa11" containerName="init" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.659339 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="462961cd-6180-4e0a-84b6-cae4c483c869" containerName="dnsmasq-dns" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.659373 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf99ed5d-244e-4006-ac58-df4cc63ffa11" containerName="dnsmasq-dns" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.660362 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.664424 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.664615 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.664694 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.668011 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.688937 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5"] Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.724170 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49fct\" (UniqueName: \"kubernetes.io/projected/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-kube-api-access-49fct\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.724595 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.724725 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.724920 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.826795 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49fct\" (UniqueName: \"kubernetes.io/projected/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-kube-api-access-49fct\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.826873 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.826910 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.826934 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.842771 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.842832 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.843168 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.847612 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49fct\" (UniqueName: \"kubernetes.io/projected/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-kube-api-access-49fct\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:27 crc kubenswrapper[4923]: I1009 10:27:27.992345 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:28 crc kubenswrapper[4923]: I1009 10:27:28.583963 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5"] Oct 09 10:27:28 crc kubenswrapper[4923]: W1009 10:27:28.586762 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69707bda_a6a2_4e7f_98c4_63a65c9d26fb.slice/crio-2470bc1e58b1b993059dfcff3becefd7dd037e521a3703758f67af9f1d14ce71 WatchSource:0}: Error finding container 2470bc1e58b1b993059dfcff3becefd7dd037e521a3703758f67af9f1d14ce71: Status 404 returned error can't find the container with id 2470bc1e58b1b993059dfcff3becefd7dd037e521a3703758f67af9f1d14ce71 Oct 09 10:27:28 crc kubenswrapper[4923]: I1009 10:27:28.638101 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" event={"ID":"69707bda-a6a2-4e7f-98c4-63a65c9d26fb","Type":"ContainerStarted","Data":"2470bc1e58b1b993059dfcff3becefd7dd037e521a3703758f67af9f1d14ce71"} Oct 09 10:27:29 crc kubenswrapper[4923]: I1009 10:27:29.654308 4923 generic.go:334] "Generic (PLEG): container finished" podID="8dbce0ef-9426-4065-9749-633e7a5e4248" containerID="02235ef01e28456a3525b1efd2f07ad05b46fb1d8584d2a2f853eebfec825646" exitCode=0 Oct 09 10:27:29 crc kubenswrapper[4923]: I1009 10:27:29.654857 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dbce0ef-9426-4065-9749-633e7a5e4248","Type":"ContainerDied","Data":"02235ef01e28456a3525b1efd2f07ad05b46fb1d8584d2a2f853eebfec825646"} Oct 09 10:27:29 crc kubenswrapper[4923]: I1009 10:27:29.662298 4923 generic.go:334] "Generic (PLEG): container finished" podID="b1b005e0-1c67-42a3-97aa-3222f4218dca" containerID="0109c70360c5b74727f7709bfb9a3d96859ab0153413ac79c881c0c607274c5f" exitCode=0 Oct 09 10:27:29 crc kubenswrapper[4923]: I1009 10:27:29.662357 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b1b005e0-1c67-42a3-97aa-3222f4218dca","Type":"ContainerDied","Data":"0109c70360c5b74727f7709bfb9a3d96859ab0153413ac79c881c0c607274c5f"} Oct 09 10:27:30 crc kubenswrapper[4923]: I1009 10:27:30.690081 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"b1b005e0-1c67-42a3-97aa-3222f4218dca","Type":"ContainerStarted","Data":"35ec0169a16430e1bbf22b0229f4367b6c49c2c72ed5628a888149752a9ec4df"} Oct 09 10:27:30 crc kubenswrapper[4923]: I1009 10:27:30.691179 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:27:30 crc kubenswrapper[4923]: I1009 10:27:30.702564 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"8dbce0ef-9426-4065-9749-633e7a5e4248","Type":"ContainerStarted","Data":"051b18b30230c94547ef905ce0828a935549799f9a153809aa3e7e63709ebe27"} Oct 09 10:27:30 crc kubenswrapper[4923]: I1009 10:27:30.703015 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 09 10:27:30 crc kubenswrapper[4923]: I1009 10:27:30.725944 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.725916776 podStartE2EDuration="36.725916776s" podCreationTimestamp="2025-10-09 10:26:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:27:30.722264316 +0000 UTC m=+1336.790446082" watchObservedRunningTime="2025-10-09 10:27:30.725916776 +0000 UTC m=+1336.794098532" Oct 09 10:27:30 crc kubenswrapper[4923]: I1009 10:27:30.766148 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.766117238 podStartE2EDuration="36.766117238s" podCreationTimestamp="2025-10-09 10:26:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 10:27:30.751788235 +0000 UTC m=+1336.819970011" watchObservedRunningTime="2025-10-09 10:27:30.766117238 +0000 UTC m=+1336.834298994" Oct 09 10:27:38 crc kubenswrapper[4923]: I1009 10:27:38.810399 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" event={"ID":"69707bda-a6a2-4e7f-98c4-63a65c9d26fb","Type":"ContainerStarted","Data":"4f2c0b6cf0546439a2a9fc4a64379397479861e1adeec4afff360b6186dc6fa5"} Oct 09 10:27:38 crc kubenswrapper[4923]: I1009 10:27:38.834812 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" podStartSLOduration=2.141504014 podStartE2EDuration="11.834783066s" podCreationTimestamp="2025-10-09 10:27:27 +0000 UTC" firstStartedPulling="2025-10-09 10:27:28.590413928 +0000 UTC m=+1334.658595694" lastFinishedPulling="2025-10-09 10:27:38.28369299 +0000 UTC m=+1344.351874746" observedRunningTime="2025-10-09 10:27:38.829063289 +0000 UTC m=+1344.897245045" watchObservedRunningTime="2025-10-09 10:27:38.834783066 +0000 UTC m=+1344.902964822" Oct 09 10:27:44 crc kubenswrapper[4923]: I1009 10:27:44.614948 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 09 10:27:44 crc kubenswrapper[4923]: I1009 10:27:44.963061 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 09 10:27:57 crc kubenswrapper[4923]: I1009 10:27:57.011476 4923 generic.go:334] "Generic (PLEG): container finished" podID="69707bda-a6a2-4e7f-98c4-63a65c9d26fb" containerID="4f2c0b6cf0546439a2a9fc4a64379397479861e1adeec4afff360b6186dc6fa5" exitCode=0 Oct 09 10:27:57 crc kubenswrapper[4923]: I1009 10:27:57.011570 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" event={"ID":"69707bda-a6a2-4e7f-98c4-63a65c9d26fb","Type":"ContainerDied","Data":"4f2c0b6cf0546439a2a9fc4a64379397479861e1adeec4afff360b6186dc6fa5"} Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.489047 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.566609 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-ssh-key\") pod \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.566699 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-repo-setup-combined-ca-bundle\") pod \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.566955 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49fct\" (UniqueName: \"kubernetes.io/projected/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-kube-api-access-49fct\") pod \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.567016 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-inventory\") pod \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\" (UID: \"69707bda-a6a2-4e7f-98c4-63a65c9d26fb\") " Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.574414 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "69707bda-a6a2-4e7f-98c4-63a65c9d26fb" (UID: "69707bda-a6a2-4e7f-98c4-63a65c9d26fb"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.575285 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-kube-api-access-49fct" (OuterVolumeSpecName: "kube-api-access-49fct") pod "69707bda-a6a2-4e7f-98c4-63a65c9d26fb" (UID: "69707bda-a6a2-4e7f-98c4-63a65c9d26fb"). InnerVolumeSpecName "kube-api-access-49fct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.600253 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-inventory" (OuterVolumeSpecName: "inventory") pod "69707bda-a6a2-4e7f-98c4-63a65c9d26fb" (UID: "69707bda-a6a2-4e7f-98c4-63a65c9d26fb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.608737 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "69707bda-a6a2-4e7f-98c4-63a65c9d26fb" (UID: "69707bda-a6a2-4e7f-98c4-63a65c9d26fb"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.670313 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49fct\" (UniqueName: \"kubernetes.io/projected/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-kube-api-access-49fct\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.670359 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.670370 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:58 crc kubenswrapper[4923]: I1009 10:27:58.670380 4923 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69707bda-a6a2-4e7f-98c4-63a65c9d26fb-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.032432 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" event={"ID":"69707bda-a6a2-4e7f-98c4-63a65c9d26fb","Type":"ContainerDied","Data":"2470bc1e58b1b993059dfcff3becefd7dd037e521a3703758f67af9f1d14ce71"} Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.032899 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2470bc1e58b1b993059dfcff3becefd7dd037e521a3703758f67af9f1d14ce71" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.032542 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.131998 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf"] Oct 09 10:27:59 crc kubenswrapper[4923]: E1009 10:27:59.132809 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69707bda-a6a2-4e7f-98c4-63a65c9d26fb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.132845 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="69707bda-a6a2-4e7f-98c4-63a65c9d26fb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.133219 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="69707bda-a6a2-4e7f-98c4-63a65c9d26fb" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.134484 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.140742 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.141067 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.141891 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.141958 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.144316 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf"] Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.179763 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.179822 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.179979 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.180220 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx8rl\" (UniqueName: \"kubernetes.io/projected/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-kube-api-access-gx8rl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.282918 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.282995 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.283039 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.283075 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx8rl\" (UniqueName: \"kubernetes.io/projected/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-kube-api-access-gx8rl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.288648 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.288648 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.290170 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.304441 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx8rl\" (UniqueName: \"kubernetes.io/projected/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-kube-api-access-gx8rl\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:27:59 crc kubenswrapper[4923]: I1009 10:27:59.464070 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:28:00 crc kubenswrapper[4923]: I1009 10:28:00.049374 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf"] Oct 09 10:28:01 crc kubenswrapper[4923]: I1009 10:28:01.057075 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" event={"ID":"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58","Type":"ContainerStarted","Data":"5bb423e27afece4cbcc63d052ac9dd55f83ac8f57afea95e5297dbd11df9257e"} Oct 09 10:28:01 crc kubenswrapper[4923]: I1009 10:28:01.057956 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" event={"ID":"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58","Type":"ContainerStarted","Data":"f5e0a7534c04377c567d963251c0203abf3e65495bb992efe5cbfa8b2b31fde6"} Oct 09 10:28:01 crc kubenswrapper[4923]: I1009 10:28:01.079807 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" podStartSLOduration=1.483860226 podStartE2EDuration="2.079776642s" podCreationTimestamp="2025-10-09 10:27:59 +0000 UTC" firstStartedPulling="2025-10-09 10:28:00.055802757 +0000 UTC m=+1366.123984513" lastFinishedPulling="2025-10-09 10:28:00.651719173 +0000 UTC m=+1366.719900929" observedRunningTime="2025-10-09 10:28:01.076546053 +0000 UTC m=+1367.144727809" watchObservedRunningTime="2025-10-09 10:28:01.079776642 +0000 UTC m=+1367.147958398" Oct 09 10:28:16 crc kubenswrapper[4923]: I1009 10:28:16.600101 4923 scope.go:117] "RemoveContainer" containerID="0720ad19d7d11b8e1b05aba153f1251872269ad2483f37c696bb36df0fa2f32b" Oct 09 10:28:16 crc kubenswrapper[4923]: I1009 10:28:16.636573 4923 scope.go:117] "RemoveContainer" containerID="238b792fa9e2d3b834704439958b1cd2c20b1c82c8d08987fd20821005b85936" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.066904 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ns95h"] Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.069819 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.086967 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ns95h"] Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.176415 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-utilities\") pod \"redhat-marketplace-ns95h\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.176624 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx5nl\" (UniqueName: \"kubernetes.io/projected/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-kube-api-access-vx5nl\") pod \"redhat-marketplace-ns95h\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.176673 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-catalog-content\") pod \"redhat-marketplace-ns95h\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.279571 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx5nl\" (UniqueName: \"kubernetes.io/projected/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-kube-api-access-vx5nl\") pod \"redhat-marketplace-ns95h\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.279674 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-catalog-content\") pod \"redhat-marketplace-ns95h\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.279925 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-utilities\") pod \"redhat-marketplace-ns95h\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.280460 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-catalog-content\") pod \"redhat-marketplace-ns95h\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.280557 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-utilities\") pod \"redhat-marketplace-ns95h\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.306841 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx5nl\" (UniqueName: \"kubernetes.io/projected/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-kube-api-access-vx5nl\") pod \"redhat-marketplace-ns95h\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.399086 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:05 crc kubenswrapper[4923]: I1009 10:29:05.919906 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ns95h"] Oct 09 10:29:06 crc kubenswrapper[4923]: I1009 10:29:06.756689 4923 generic.go:334] "Generic (PLEG): container finished" podID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerID="6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4" exitCode=0 Oct 09 10:29:06 crc kubenswrapper[4923]: I1009 10:29:06.757199 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ns95h" event={"ID":"30c06e32-18f5-4e66-bce6-fe7bb46be5ad","Type":"ContainerDied","Data":"6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4"} Oct 09 10:29:06 crc kubenswrapper[4923]: I1009 10:29:06.757238 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ns95h" event={"ID":"30c06e32-18f5-4e66-bce6-fe7bb46be5ad","Type":"ContainerStarted","Data":"adb43c15f953b7fb47f5e05a05996ce5191cfea9e2decf7af3e95dddacb77fe6"} Oct 09 10:29:08 crc kubenswrapper[4923]: I1009 10:29:08.780470 4923 generic.go:334] "Generic (PLEG): container finished" podID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerID="230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84" exitCode=0 Oct 09 10:29:08 crc kubenswrapper[4923]: I1009 10:29:08.780604 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ns95h" event={"ID":"30c06e32-18f5-4e66-bce6-fe7bb46be5ad","Type":"ContainerDied","Data":"230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84"} Oct 09 10:29:09 crc kubenswrapper[4923]: I1009 10:29:09.813326 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ns95h" event={"ID":"30c06e32-18f5-4e66-bce6-fe7bb46be5ad","Type":"ContainerStarted","Data":"657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee"} Oct 09 10:29:09 crc kubenswrapper[4923]: I1009 10:29:09.858746 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ns95h" podStartSLOduration=2.3021100199999998 podStartE2EDuration="4.858717122s" podCreationTimestamp="2025-10-09 10:29:05 +0000 UTC" firstStartedPulling="2025-10-09 10:29:06.759338045 +0000 UTC m=+1432.827519811" lastFinishedPulling="2025-10-09 10:29:09.315945167 +0000 UTC m=+1435.384126913" observedRunningTime="2025-10-09 10:29:09.838862702 +0000 UTC m=+1435.907044468" watchObservedRunningTime="2025-10-09 10:29:09.858717122 +0000 UTC m=+1435.926898888" Oct 09 10:29:15 crc kubenswrapper[4923]: I1009 10:29:15.399607 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:15 crc kubenswrapper[4923]: I1009 10:29:15.400069 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:15 crc kubenswrapper[4923]: I1009 10:29:15.453005 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:15 crc kubenswrapper[4923]: I1009 10:29:15.933280 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:15 crc kubenswrapper[4923]: I1009 10:29:15.993435 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ns95h"] Oct 09 10:29:16 crc kubenswrapper[4923]: I1009 10:29:16.741123 4923 scope.go:117] "RemoveContainer" containerID="3ce6673e43a0bde733503e5d3e1e3534302cb96687208e1450fb761722773acb" Oct 09 10:29:17 crc kubenswrapper[4923]: I1009 10:29:17.902675 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ns95h" podUID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerName="registry-server" containerID="cri-o://657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee" gracePeriod=2 Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.366874 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.477282 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-utilities\") pod \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.477479 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-catalog-content\") pod \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.477775 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vx5nl\" (UniqueName: \"kubernetes.io/projected/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-kube-api-access-vx5nl\") pod \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\" (UID: \"30c06e32-18f5-4e66-bce6-fe7bb46be5ad\") " Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.478514 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-utilities" (OuterVolumeSpecName: "utilities") pod "30c06e32-18f5-4e66-bce6-fe7bb46be5ad" (UID: "30c06e32-18f5-4e66-bce6-fe7bb46be5ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.487172 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-kube-api-access-vx5nl" (OuterVolumeSpecName: "kube-api-access-vx5nl") pod "30c06e32-18f5-4e66-bce6-fe7bb46be5ad" (UID: "30c06e32-18f5-4e66-bce6-fe7bb46be5ad"). InnerVolumeSpecName "kube-api-access-vx5nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.493575 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "30c06e32-18f5-4e66-bce6-fe7bb46be5ad" (UID: "30c06e32-18f5-4e66-bce6-fe7bb46be5ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.580451 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.580508 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.580521 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vx5nl\" (UniqueName: \"kubernetes.io/projected/30c06e32-18f5-4e66-bce6-fe7bb46be5ad-kube-api-access-vx5nl\") on node \"crc\" DevicePath \"\"" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.913788 4923 generic.go:334] "Generic (PLEG): container finished" podID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerID="657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee" exitCode=0 Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.913854 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ns95h" event={"ID":"30c06e32-18f5-4e66-bce6-fe7bb46be5ad","Type":"ContainerDied","Data":"657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee"} Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.913890 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ns95h" event={"ID":"30c06e32-18f5-4e66-bce6-fe7bb46be5ad","Type":"ContainerDied","Data":"adb43c15f953b7fb47f5e05a05996ce5191cfea9e2decf7af3e95dddacb77fe6"} Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.913914 4923 scope.go:117] "RemoveContainer" containerID="657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.914065 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ns95h" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.941308 4923 scope.go:117] "RemoveContainer" containerID="230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84" Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.959311 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ns95h"] Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.968742 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ns95h"] Oct 09 10:29:18 crc kubenswrapper[4923]: I1009 10:29:18.975461 4923 scope.go:117] "RemoveContainer" containerID="6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4" Oct 09 10:29:19 crc kubenswrapper[4923]: I1009 10:29:19.031343 4923 scope.go:117] "RemoveContainer" containerID="657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee" Oct 09 10:29:19 crc kubenswrapper[4923]: E1009 10:29:19.032002 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee\": container with ID starting with 657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee not found: ID does not exist" containerID="657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee" Oct 09 10:29:19 crc kubenswrapper[4923]: I1009 10:29:19.032060 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee"} err="failed to get container status \"657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee\": rpc error: code = NotFound desc = could not find container \"657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee\": container with ID starting with 657740e0435b47549648fd02d6f61bfb00c569e4a42f8bad2756e9b30afedeee not found: ID does not exist" Oct 09 10:29:19 crc kubenswrapper[4923]: I1009 10:29:19.032149 4923 scope.go:117] "RemoveContainer" containerID="230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84" Oct 09 10:29:19 crc kubenswrapper[4923]: E1009 10:29:19.032676 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84\": container with ID starting with 230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84 not found: ID does not exist" containerID="230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84" Oct 09 10:29:19 crc kubenswrapper[4923]: I1009 10:29:19.032868 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84"} err="failed to get container status \"230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84\": rpc error: code = NotFound desc = could not find container \"230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84\": container with ID starting with 230a68f5d864b32f527292e2c0ecbdf84f4f30d58be4c2369fcd39b67a740c84 not found: ID does not exist" Oct 09 10:29:19 crc kubenswrapper[4923]: I1009 10:29:19.032906 4923 scope.go:117] "RemoveContainer" containerID="6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4" Oct 09 10:29:19 crc kubenswrapper[4923]: E1009 10:29:19.033491 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4\": container with ID starting with 6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4 not found: ID does not exist" containerID="6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4" Oct 09 10:29:19 crc kubenswrapper[4923]: I1009 10:29:19.033531 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4"} err="failed to get container status \"6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4\": rpc error: code = NotFound desc = could not find container \"6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4\": container with ID starting with 6e30b383e29a2fe504f68d98f715bcd075684acf5822005c499abcb562ee6df4 not found: ID does not exist" Oct 09 10:29:20 crc kubenswrapper[4923]: I1009 10:29:20.615044 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" path="/var/lib/kubelet/pods/30c06e32-18f5-4e66-bce6-fe7bb46be5ad/volumes" Oct 09 10:29:24 crc kubenswrapper[4923]: I1009 10:29:24.600051 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:29:24 crc kubenswrapper[4923]: I1009 10:29:24.600447 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.281028 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ssm8q"] Oct 09 10:29:47 crc kubenswrapper[4923]: E1009 10:29:47.282664 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerName="registry-server" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.282685 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerName="registry-server" Oct 09 10:29:47 crc kubenswrapper[4923]: E1009 10:29:47.282717 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerName="extract-utilities" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.282726 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerName="extract-utilities" Oct 09 10:29:47 crc kubenswrapper[4923]: E1009 10:29:47.282771 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerName="extract-content" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.282781 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerName="extract-content" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.283063 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="30c06e32-18f5-4e66-bce6-fe7bb46be5ad" containerName="registry-server" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.285171 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.312482 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ssm8q"] Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.339658 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-catalog-content\") pod \"community-operators-ssm8q\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.339735 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-utilities\") pod \"community-operators-ssm8q\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.339883 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d5f5\" (UniqueName: \"kubernetes.io/projected/c1a01e69-cb11-4e95-abc0-9995d5189478-kube-api-access-8d5f5\") pod \"community-operators-ssm8q\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.442536 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-utilities\") pod \"community-operators-ssm8q\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.442600 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d5f5\" (UniqueName: \"kubernetes.io/projected/c1a01e69-cb11-4e95-abc0-9995d5189478-kube-api-access-8d5f5\") pod \"community-operators-ssm8q\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.442852 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-catalog-content\") pod \"community-operators-ssm8q\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.443245 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-catalog-content\") pod \"community-operators-ssm8q\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.443286 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-utilities\") pod \"community-operators-ssm8q\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.469786 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d5f5\" (UniqueName: \"kubernetes.io/projected/c1a01e69-cb11-4e95-abc0-9995d5189478-kube-api-access-8d5f5\") pod \"community-operators-ssm8q\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:47 crc kubenswrapper[4923]: I1009 10:29:47.616289 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:48 crc kubenswrapper[4923]: I1009 10:29:48.192970 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ssm8q"] Oct 09 10:29:48 crc kubenswrapper[4923]: I1009 10:29:48.233587 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssm8q" event={"ID":"c1a01e69-cb11-4e95-abc0-9995d5189478","Type":"ContainerStarted","Data":"5811c86f2b2c30be7771ba99b6339c073de90311f244e1ffbe08c424a0dd24bd"} Oct 09 10:29:49 crc kubenswrapper[4923]: I1009 10:29:49.246922 4923 generic.go:334] "Generic (PLEG): container finished" podID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerID="69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2" exitCode=0 Oct 09 10:29:49 crc kubenswrapper[4923]: I1009 10:29:49.247355 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssm8q" event={"ID":"c1a01e69-cb11-4e95-abc0-9995d5189478","Type":"ContainerDied","Data":"69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2"} Oct 09 10:29:50 crc kubenswrapper[4923]: I1009 10:29:50.283709 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssm8q" event={"ID":"c1a01e69-cb11-4e95-abc0-9995d5189478","Type":"ContainerStarted","Data":"c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73"} Oct 09 10:29:51 crc kubenswrapper[4923]: I1009 10:29:51.295522 4923 generic.go:334] "Generic (PLEG): container finished" podID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerID="c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73" exitCode=0 Oct 09 10:29:51 crc kubenswrapper[4923]: I1009 10:29:51.295576 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssm8q" event={"ID":"c1a01e69-cb11-4e95-abc0-9995d5189478","Type":"ContainerDied","Data":"c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73"} Oct 09 10:29:52 crc kubenswrapper[4923]: I1009 10:29:52.308668 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssm8q" event={"ID":"c1a01e69-cb11-4e95-abc0-9995d5189478","Type":"ContainerStarted","Data":"9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352"} Oct 09 10:29:52 crc kubenswrapper[4923]: I1009 10:29:52.332737 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ssm8q" podStartSLOduration=2.77069598 podStartE2EDuration="5.332706272s" podCreationTimestamp="2025-10-09 10:29:47 +0000 UTC" firstStartedPulling="2025-10-09 10:29:49.249563124 +0000 UTC m=+1475.317744890" lastFinishedPulling="2025-10-09 10:29:51.811573426 +0000 UTC m=+1477.879755182" observedRunningTime="2025-10-09 10:29:52.327574279 +0000 UTC m=+1478.395756035" watchObservedRunningTime="2025-10-09 10:29:52.332706272 +0000 UTC m=+1478.400888028" Oct 09 10:29:54 crc kubenswrapper[4923]: I1009 10:29:54.599839 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:29:54 crc kubenswrapper[4923]: I1009 10:29:54.600299 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:29:57 crc kubenswrapper[4923]: I1009 10:29:57.622706 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:57 crc kubenswrapper[4923]: I1009 10:29:57.623992 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:57 crc kubenswrapper[4923]: I1009 10:29:57.686241 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:58 crc kubenswrapper[4923]: I1009 10:29:58.425312 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:29:58 crc kubenswrapper[4923]: I1009 10:29:58.483871 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ssm8q"] Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.155450 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct"] Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.157408 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.162212 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.165717 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.171837 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct"] Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.230427 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8934ec94-c42f-4428-9f86-5b8b90ed2d56-secret-volume\") pod \"collect-profiles-29333430-mh5ct\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.230565 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbfhj\" (UniqueName: \"kubernetes.io/projected/8934ec94-c42f-4428-9f86-5b8b90ed2d56-kube-api-access-sbfhj\") pod \"collect-profiles-29333430-mh5ct\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.230621 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8934ec94-c42f-4428-9f86-5b8b90ed2d56-config-volume\") pod \"collect-profiles-29333430-mh5ct\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.333448 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8934ec94-c42f-4428-9f86-5b8b90ed2d56-secret-volume\") pod \"collect-profiles-29333430-mh5ct\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.333519 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbfhj\" (UniqueName: \"kubernetes.io/projected/8934ec94-c42f-4428-9f86-5b8b90ed2d56-kube-api-access-sbfhj\") pod \"collect-profiles-29333430-mh5ct\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.333559 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8934ec94-c42f-4428-9f86-5b8b90ed2d56-config-volume\") pod \"collect-profiles-29333430-mh5ct\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.334857 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8934ec94-c42f-4428-9f86-5b8b90ed2d56-config-volume\") pod \"collect-profiles-29333430-mh5ct\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.341059 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8934ec94-c42f-4428-9f86-5b8b90ed2d56-secret-volume\") pod \"collect-profiles-29333430-mh5ct\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.354377 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbfhj\" (UniqueName: \"kubernetes.io/projected/8934ec94-c42f-4428-9f86-5b8b90ed2d56-kube-api-access-sbfhj\") pod \"collect-profiles-29333430-mh5ct\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.401067 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ssm8q" podUID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerName="registry-server" containerID="cri-o://9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352" gracePeriod=2 Oct 09 10:30:00 crc kubenswrapper[4923]: I1009 10:30:00.489499 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:00.918611 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.050159 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-utilities\") pod \"c1a01e69-cb11-4e95-abc0-9995d5189478\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.050494 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d5f5\" (UniqueName: \"kubernetes.io/projected/c1a01e69-cb11-4e95-abc0-9995d5189478-kube-api-access-8d5f5\") pod \"c1a01e69-cb11-4e95-abc0-9995d5189478\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.050686 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-catalog-content\") pod \"c1a01e69-cb11-4e95-abc0-9995d5189478\" (UID: \"c1a01e69-cb11-4e95-abc0-9995d5189478\") " Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.052007 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-utilities" (OuterVolumeSpecName: "utilities") pod "c1a01e69-cb11-4e95-abc0-9995d5189478" (UID: "c1a01e69-cb11-4e95-abc0-9995d5189478"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.058197 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1a01e69-cb11-4e95-abc0-9995d5189478-kube-api-access-8d5f5" (OuterVolumeSpecName: "kube-api-access-8d5f5") pod "c1a01e69-cb11-4e95-abc0-9995d5189478" (UID: "c1a01e69-cb11-4e95-abc0-9995d5189478"). InnerVolumeSpecName "kube-api-access-8d5f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.061510 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct"] Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.154368 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.154415 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d5f5\" (UniqueName: \"kubernetes.io/projected/c1a01e69-cb11-4e95-abc0-9995d5189478-kube-api-access-8d5f5\") on node \"crc\" DevicePath \"\"" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.413213 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" event={"ID":"8934ec94-c42f-4428-9f86-5b8b90ed2d56","Type":"ContainerStarted","Data":"bb3c40551af9b90767eb8bb7129cdbf095121bb9e4bc8beb57953b297d51fe34"} Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.415687 4923 generic.go:334] "Generic (PLEG): container finished" podID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerID="9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352" exitCode=0 Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.415726 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssm8q" event={"ID":"c1a01e69-cb11-4e95-abc0-9995d5189478","Type":"ContainerDied","Data":"9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352"} Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.415759 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssm8q" event={"ID":"c1a01e69-cb11-4e95-abc0-9995d5189478","Type":"ContainerDied","Data":"5811c86f2b2c30be7771ba99b6339c073de90311f244e1ffbe08c424a0dd24bd"} Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.415821 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ssm8q" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.415821 4923 scope.go:117] "RemoveContainer" containerID="9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.441184 4923 scope.go:117] "RemoveContainer" containerID="c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.448153 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c1a01e69-cb11-4e95-abc0-9995d5189478" (UID: "c1a01e69-cb11-4e95-abc0-9995d5189478"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.464940 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c1a01e69-cb11-4e95-abc0-9995d5189478-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.478928 4923 scope.go:117] "RemoveContainer" containerID="69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.526690 4923 scope.go:117] "RemoveContainer" containerID="9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352" Oct 09 10:30:01 crc kubenswrapper[4923]: E1009 10:30:01.527572 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352\": container with ID starting with 9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352 not found: ID does not exist" containerID="9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.527607 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352"} err="failed to get container status \"9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352\": rpc error: code = NotFound desc = could not find container \"9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352\": container with ID starting with 9a782c421e8f5c7c514b2e67a9b7868584928658f06f6712abf650e8bb9ce352 not found: ID does not exist" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.527642 4923 scope.go:117] "RemoveContainer" containerID="c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73" Oct 09 10:30:01 crc kubenswrapper[4923]: E1009 10:30:01.528159 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73\": container with ID starting with c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73 not found: ID does not exist" containerID="c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.528205 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73"} err="failed to get container status \"c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73\": rpc error: code = NotFound desc = could not find container \"c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73\": container with ID starting with c2250faa0eb8c11c719fae9afcce624b1f1cdfcbdc27293657bec417e7463c73 not found: ID does not exist" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.528238 4923 scope.go:117] "RemoveContainer" containerID="69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2" Oct 09 10:30:01 crc kubenswrapper[4923]: E1009 10:30:01.528541 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2\": container with ID starting with 69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2 not found: ID does not exist" containerID="69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.528566 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2"} err="failed to get container status \"69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2\": rpc error: code = NotFound desc = could not find container \"69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2\": container with ID starting with 69c6c3c6647ce93bcf98b8186a6eac0797236a4f2095220d6aeccd244bed09b2 not found: ID does not exist" Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.762737 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ssm8q"] Oct 09 10:30:01 crc kubenswrapper[4923]: I1009 10:30:01.804891 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ssm8q"] Oct 09 10:30:02 crc kubenswrapper[4923]: I1009 10:30:02.428628 4923 generic.go:334] "Generic (PLEG): container finished" podID="8934ec94-c42f-4428-9f86-5b8b90ed2d56" containerID="33d99d075d4017b07613e953520de05c07eb7f7e92efa16a56c4f482c655c1c2" exitCode=0 Oct 09 10:30:02 crc kubenswrapper[4923]: I1009 10:30:02.428724 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" event={"ID":"8934ec94-c42f-4428-9f86-5b8b90ed2d56","Type":"ContainerDied","Data":"33d99d075d4017b07613e953520de05c07eb7f7e92efa16a56c4f482c655c1c2"} Oct 09 10:30:02 crc kubenswrapper[4923]: I1009 10:30:02.618634 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1a01e69-cb11-4e95-abc0-9995d5189478" path="/var/lib/kubelet/pods/c1a01e69-cb11-4e95-abc0-9995d5189478/volumes" Oct 09 10:30:03 crc kubenswrapper[4923]: I1009 10:30:03.781703 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:03 crc kubenswrapper[4923]: I1009 10:30:03.920130 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbfhj\" (UniqueName: \"kubernetes.io/projected/8934ec94-c42f-4428-9f86-5b8b90ed2d56-kube-api-access-sbfhj\") pod \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " Oct 09 10:30:03 crc kubenswrapper[4923]: I1009 10:30:03.920322 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8934ec94-c42f-4428-9f86-5b8b90ed2d56-secret-volume\") pod \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " Oct 09 10:30:03 crc kubenswrapper[4923]: I1009 10:30:03.920405 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8934ec94-c42f-4428-9f86-5b8b90ed2d56-config-volume\") pod \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\" (UID: \"8934ec94-c42f-4428-9f86-5b8b90ed2d56\") " Oct 09 10:30:03 crc kubenswrapper[4923]: I1009 10:30:03.921300 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8934ec94-c42f-4428-9f86-5b8b90ed2d56-config-volume" (OuterVolumeSpecName: "config-volume") pod "8934ec94-c42f-4428-9f86-5b8b90ed2d56" (UID: "8934ec94-c42f-4428-9f86-5b8b90ed2d56"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:30:03 crc kubenswrapper[4923]: I1009 10:30:03.928288 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8934ec94-c42f-4428-9f86-5b8b90ed2d56-kube-api-access-sbfhj" (OuterVolumeSpecName: "kube-api-access-sbfhj") pod "8934ec94-c42f-4428-9f86-5b8b90ed2d56" (UID: "8934ec94-c42f-4428-9f86-5b8b90ed2d56"). InnerVolumeSpecName "kube-api-access-sbfhj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:30:03 crc kubenswrapper[4923]: I1009 10:30:03.928690 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8934ec94-c42f-4428-9f86-5b8b90ed2d56-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8934ec94-c42f-4428-9f86-5b8b90ed2d56" (UID: "8934ec94-c42f-4428-9f86-5b8b90ed2d56"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:30:04 crc kubenswrapper[4923]: I1009 10:30:04.024580 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8934ec94-c42f-4428-9f86-5b8b90ed2d56-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 10:30:04 crc kubenswrapper[4923]: I1009 10:30:04.024651 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbfhj\" (UniqueName: \"kubernetes.io/projected/8934ec94-c42f-4428-9f86-5b8b90ed2d56-kube-api-access-sbfhj\") on node \"crc\" DevicePath \"\"" Oct 09 10:30:04 crc kubenswrapper[4923]: I1009 10:30:04.024669 4923 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8934ec94-c42f-4428-9f86-5b8b90ed2d56-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 10:30:04 crc kubenswrapper[4923]: I1009 10:30:04.450721 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" event={"ID":"8934ec94-c42f-4428-9f86-5b8b90ed2d56","Type":"ContainerDied","Data":"bb3c40551af9b90767eb8bb7129cdbf095121bb9e4bc8beb57953b297d51fe34"} Oct 09 10:30:04 crc kubenswrapper[4923]: I1009 10:30:04.450857 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333430-mh5ct" Oct 09 10:30:04 crc kubenswrapper[4923]: I1009 10:30:04.450866 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb3c40551af9b90767eb8bb7129cdbf095121bb9e4bc8beb57953b297d51fe34" Oct 09 10:30:24 crc kubenswrapper[4923]: I1009 10:30:24.600397 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:30:24 crc kubenswrapper[4923]: I1009 10:30:24.601370 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:30:24 crc kubenswrapper[4923]: I1009 10:30:24.726266 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:30:24 crc kubenswrapper[4923]: I1009 10:30:24.743256 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:30:24 crc kubenswrapper[4923]: I1009 10:30:24.743413 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" gracePeriod=600 Oct 09 10:30:24 crc kubenswrapper[4923]: E1009 10:30:24.905400 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:30:25 crc kubenswrapper[4923]: I1009 10:30:25.735652 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" exitCode=0 Oct 09 10:30:25 crc kubenswrapper[4923]: I1009 10:30:25.735735 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487"} Oct 09 10:30:25 crc kubenswrapper[4923]: I1009 10:30:25.735836 4923 scope.go:117] "RemoveContainer" containerID="cb8062901f893bbba9ad244080627bdc62960c7f3035d2ee9483ebb1559dcc72" Oct 09 10:30:25 crc kubenswrapper[4923]: I1009 10:30:25.736630 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:30:25 crc kubenswrapper[4923]: E1009 10:30:25.736994 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:30:36 crc kubenswrapper[4923]: I1009 10:30:36.601262 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:30:36 crc kubenswrapper[4923]: E1009 10:30:36.602356 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:30:51 crc kubenswrapper[4923]: I1009 10:30:51.602381 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:30:51 crc kubenswrapper[4923]: E1009 10:30:51.603801 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:31:02 crc kubenswrapper[4923]: I1009 10:31:02.602134 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:31:02 crc kubenswrapper[4923]: E1009 10:31:02.603057 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:31:14 crc kubenswrapper[4923]: I1009 10:31:14.610143 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:31:14 crc kubenswrapper[4923]: E1009 10:31:14.613395 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:31:16 crc kubenswrapper[4923]: I1009 10:31:16.887171 4923 scope.go:117] "RemoveContainer" containerID="40fc71db54b599f418267f1ad1da8f30e84e7b8de3bc8701928ee0228a83ba54" Oct 09 10:31:16 crc kubenswrapper[4923]: I1009 10:31:16.942996 4923 scope.go:117] "RemoveContainer" containerID="19ced66d3a8413134b8e05e7b359f29a4b16c6b7ce0d1861cb75903f98fe986b" Oct 09 10:31:25 crc kubenswrapper[4923]: I1009 10:31:25.044628 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-679ld"] Oct 09 10:31:25 crc kubenswrapper[4923]: I1009 10:31:25.054866 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-vxmdr"] Oct 09 10:31:25 crc kubenswrapper[4923]: I1009 10:31:25.064730 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-vxmdr"] Oct 09 10:31:25 crc kubenswrapper[4923]: I1009 10:31:25.076436 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-679ld"] Oct 09 10:31:26 crc kubenswrapper[4923]: I1009 10:31:26.032376 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-sbbl8"] Oct 09 10:31:26 crc kubenswrapper[4923]: I1009 10:31:26.041566 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-sbbl8"] Oct 09 10:31:26 crc kubenswrapper[4923]: I1009 10:31:26.615598 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4e8ce207-1c54-43ea-8f35-a1a47088d249" path="/var/lib/kubelet/pods/4e8ce207-1c54-43ea-8f35-a1a47088d249/volumes" Oct 09 10:31:26 crc kubenswrapper[4923]: I1009 10:31:26.617148 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3b07ecd-e8d3-494a-aefe-5c8204428bb6" path="/var/lib/kubelet/pods/d3b07ecd-e8d3-494a-aefe-5c8204428bb6/volumes" Oct 09 10:31:26 crc kubenswrapper[4923]: I1009 10:31:26.617853 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1f95ef9-7942-49c4-971b-9871c3a84b37" path="/var/lib/kubelet/pods/e1f95ef9-7942-49c4-971b-9871c3a84b37/volumes" Oct 09 10:31:27 crc kubenswrapper[4923]: I1009 10:31:27.602287 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:31:27 crc kubenswrapper[4923]: E1009 10:31:27.603129 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:31:35 crc kubenswrapper[4923]: I1009 10:31:35.033783 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-7825-account-create-phb7p"] Oct 09 10:31:35 crc kubenswrapper[4923]: I1009 10:31:35.044799 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7555-account-create-rqrzd"] Oct 09 10:31:35 crc kubenswrapper[4923]: I1009 10:31:35.055823 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-7825-account-create-phb7p"] Oct 09 10:31:35 crc kubenswrapper[4923]: I1009 10:31:35.064369 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7555-account-create-rqrzd"] Oct 09 10:31:36 crc kubenswrapper[4923]: I1009 10:31:36.055430 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-be11-account-create-9ms9b"] Oct 09 10:31:36 crc kubenswrapper[4923]: I1009 10:31:36.082413 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-be11-account-create-9ms9b"] Oct 09 10:31:36 crc kubenswrapper[4923]: I1009 10:31:36.614380 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31ff6e52-d758-44a7-801c-7a0a9578157a" path="/var/lib/kubelet/pods/31ff6e52-d758-44a7-801c-7a0a9578157a/volumes" Oct 09 10:31:36 crc kubenswrapper[4923]: I1009 10:31:36.615830 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="835679c7-23a5-4347-b57c-eb0d600195b4" path="/var/lib/kubelet/pods/835679c7-23a5-4347-b57c-eb0d600195b4/volumes" Oct 09 10:31:36 crc kubenswrapper[4923]: I1009 10:31:36.616573 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0" path="/var/lib/kubelet/pods/e5eb6fa4-e4e1-4909-a2f4-9870099f7eb0/volumes" Oct 09 10:31:42 crc kubenswrapper[4923]: I1009 10:31:42.601460 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:31:42 crc kubenswrapper[4923]: E1009 10:31:42.602707 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:31:55 crc kubenswrapper[4923]: I1009 10:31:55.031637 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-bwv6q"] Oct 09 10:31:55 crc kubenswrapper[4923]: I1009 10:31:55.042006 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-p77wh"] Oct 09 10:31:55 crc kubenswrapper[4923]: I1009 10:31:55.052160 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-4545q"] Oct 09 10:31:55 crc kubenswrapper[4923]: I1009 10:31:55.064693 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-4545q"] Oct 09 10:31:55 crc kubenswrapper[4923]: I1009 10:31:55.078101 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-bwv6q"] Oct 09 10:31:55 crc kubenswrapper[4923]: I1009 10:31:55.097376 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-p77wh"] Oct 09 10:31:55 crc kubenswrapper[4923]: I1009 10:31:55.602214 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:31:55 crc kubenswrapper[4923]: E1009 10:31:55.602540 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:31:55 crc kubenswrapper[4923]: I1009 10:31:55.644129 4923 generic.go:334] "Generic (PLEG): container finished" podID="1d901fa9-21b4-4b52-bbd1-11aa7cd06e58" containerID="5bb423e27afece4cbcc63d052ac9dd55f83ac8f57afea95e5297dbd11df9257e" exitCode=0 Oct 09 10:31:55 crc kubenswrapper[4923]: I1009 10:31:55.644181 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" event={"ID":"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58","Type":"ContainerDied","Data":"5bb423e27afece4cbcc63d052ac9dd55f83ac8f57afea95e5297dbd11df9257e"} Oct 09 10:31:56 crc kubenswrapper[4923]: I1009 10:31:56.615954 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e7d03c1-0cd8-41ca-aba6-880cf6fc2674" path="/var/lib/kubelet/pods/5e7d03c1-0cd8-41ca-aba6-880cf6fc2674/volumes" Oct 09 10:31:56 crc kubenswrapper[4923]: I1009 10:31:56.616983 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="629aac6b-a0f1-4a58-b491-489a882cf65c" path="/var/lib/kubelet/pods/629aac6b-a0f1-4a58-b491-489a882cf65c/volumes" Oct 09 10:31:56 crc kubenswrapper[4923]: I1009 10:31:56.617481 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1a8663c-24f3-4917-b301-4880439294df" path="/var/lib/kubelet/pods/a1a8663c-24f3-4917-b301-4880439294df/volumes" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.089245 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.209767 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gx8rl\" (UniqueName: \"kubernetes.io/projected/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-kube-api-access-gx8rl\") pod \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.209834 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-inventory\") pod \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.209861 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-ssh-key\") pod \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.209955 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-bootstrap-combined-ca-bundle\") pod \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\" (UID: \"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58\") " Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.216321 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-kube-api-access-gx8rl" (OuterVolumeSpecName: "kube-api-access-gx8rl") pod "1d901fa9-21b4-4b52-bbd1-11aa7cd06e58" (UID: "1d901fa9-21b4-4b52-bbd1-11aa7cd06e58"). InnerVolumeSpecName "kube-api-access-gx8rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.216321 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1d901fa9-21b4-4b52-bbd1-11aa7cd06e58" (UID: "1d901fa9-21b4-4b52-bbd1-11aa7cd06e58"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.243272 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-inventory" (OuterVolumeSpecName: "inventory") pod "1d901fa9-21b4-4b52-bbd1-11aa7cd06e58" (UID: "1d901fa9-21b4-4b52-bbd1-11aa7cd06e58"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.243717 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1d901fa9-21b4-4b52-bbd1-11aa7cd06e58" (UID: "1d901fa9-21b4-4b52-bbd1-11aa7cd06e58"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.312024 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gx8rl\" (UniqueName: \"kubernetes.io/projected/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-kube-api-access-gx8rl\") on node \"crc\" DevicePath \"\"" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.312065 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.312075 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.312089 4923 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.667150 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" event={"ID":"1d901fa9-21b4-4b52-bbd1-11aa7cd06e58","Type":"ContainerDied","Data":"f5e0a7534c04377c567d963251c0203abf3e65495bb992efe5cbfa8b2b31fde6"} Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.667209 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5e0a7534c04377c567d963251c0203abf3e65495bb992efe5cbfa8b2b31fde6" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.667231 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.776633 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7"] Oct 09 10:31:57 crc kubenswrapper[4923]: E1009 10:31:57.777433 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8934ec94-c42f-4428-9f86-5b8b90ed2d56" containerName="collect-profiles" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.777463 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8934ec94-c42f-4428-9f86-5b8b90ed2d56" containerName="collect-profiles" Oct 09 10:31:57 crc kubenswrapper[4923]: E1009 10:31:57.777515 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerName="extract-utilities" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.777529 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerName="extract-utilities" Oct 09 10:31:57 crc kubenswrapper[4923]: E1009 10:31:57.777538 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerName="registry-server" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.777547 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerName="registry-server" Oct 09 10:31:57 crc kubenswrapper[4923]: E1009 10:31:57.777572 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerName="extract-content" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.777579 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerName="extract-content" Oct 09 10:31:57 crc kubenswrapper[4923]: E1009 10:31:57.777593 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d901fa9-21b4-4b52-bbd1-11aa7cd06e58" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.777601 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d901fa9-21b4-4b52-bbd1-11aa7cd06e58" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.778014 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="8934ec94-c42f-4428-9f86-5b8b90ed2d56" containerName="collect-profiles" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.778052 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1a01e69-cb11-4e95-abc0-9995d5189478" containerName="registry-server" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.778063 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d901fa9-21b4-4b52-bbd1-11aa7cd06e58" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.778948 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.783319 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.783980 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.784042 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.783985 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.789891 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7"] Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.924602 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.924697 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:57 crc kubenswrapper[4923]: I1009 10:31:57.924840 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwrgt\" (UniqueName: \"kubernetes.io/projected/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-kube-api-access-cwrgt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:58 crc kubenswrapper[4923]: I1009 10:31:58.026478 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:58 crc kubenswrapper[4923]: I1009 10:31:58.026536 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:58 crc kubenswrapper[4923]: I1009 10:31:58.026593 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwrgt\" (UniqueName: \"kubernetes.io/projected/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-kube-api-access-cwrgt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:58 crc kubenswrapper[4923]: I1009 10:31:58.032729 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:58 crc kubenswrapper[4923]: I1009 10:31:58.036271 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:58 crc kubenswrapper[4923]: I1009 10:31:58.046923 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwrgt\" (UniqueName: \"kubernetes.io/projected/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-kube-api-access-cwrgt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:58 crc kubenswrapper[4923]: I1009 10:31:58.096926 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:31:58 crc kubenswrapper[4923]: I1009 10:31:58.736695 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7"] Oct 09 10:31:58 crc kubenswrapper[4923]: I1009 10:31:58.738961 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 10:31:59 crc kubenswrapper[4923]: I1009 10:31:59.685660 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" event={"ID":"00251bc8-62c6-4f4e-8aa0-f8da07570f3a","Type":"ContainerStarted","Data":"f008a5c2a03529e245025565ddfe6ff35b0945349b0d044dd926bffbaf5cefa9"} Oct 09 10:32:00 crc kubenswrapper[4923]: I1009 10:32:00.697015 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" event={"ID":"00251bc8-62c6-4f4e-8aa0-f8da07570f3a","Type":"ContainerStarted","Data":"6f0cd671f826c9d29056422776a0cf628ae70499aefc6b58dccc8b71acb825ea"} Oct 09 10:32:00 crc kubenswrapper[4923]: I1009 10:32:00.722458 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" podStartSLOduration=2.916749658 podStartE2EDuration="3.722436461s" podCreationTimestamp="2025-10-09 10:31:57 +0000 UTC" firstStartedPulling="2025-10-09 10:31:58.738647606 +0000 UTC m=+1604.806829362" lastFinishedPulling="2025-10-09 10:31:59.544334389 +0000 UTC m=+1605.612516165" observedRunningTime="2025-10-09 10:32:00.716989551 +0000 UTC m=+1606.785171337" watchObservedRunningTime="2025-10-09 10:32:00.722436461 +0000 UTC m=+1606.790618217" Oct 09 10:32:08 crc kubenswrapper[4923]: I1009 10:32:08.051253 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-rmkl7"] Oct 09 10:32:08 crc kubenswrapper[4923]: I1009 10:32:08.064289 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-rmkl7"] Oct 09 10:32:08 crc kubenswrapper[4923]: I1009 10:32:08.617580 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8febaa7-0d33-4ec9-90ce-07f415538366" path="/var/lib/kubelet/pods/f8febaa7-0d33-4ec9-90ce-07f415538366/volumes" Oct 09 10:32:09 crc kubenswrapper[4923]: I1009 10:32:09.601454 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:32:09 crc kubenswrapper[4923]: E1009 10:32:09.602192 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:32:12 crc kubenswrapper[4923]: I1009 10:32:12.033225 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-51be-account-create-dswv4"] Oct 09 10:32:12 crc kubenswrapper[4923]: I1009 10:32:12.042007 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-51be-account-create-dswv4"] Oct 09 10:32:12 crc kubenswrapper[4923]: I1009 10:32:12.637812 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3151f13-c711-4e82-b24b-53be5ecf3e66" path="/var/lib/kubelet/pods/d3151f13-c711-4e82-b24b-53be5ecf3e66/volumes" Oct 09 10:32:14 crc kubenswrapper[4923]: I1009 10:32:14.041781 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-55d5-account-create-wsrr5"] Oct 09 10:32:14 crc kubenswrapper[4923]: I1009 10:32:14.053678 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-3a12-account-create-wsttd"] Oct 09 10:32:14 crc kubenswrapper[4923]: I1009 10:32:14.064640 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-3a12-account-create-wsttd"] Oct 09 10:32:14 crc kubenswrapper[4923]: I1009 10:32:14.074869 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-55d5-account-create-wsrr5"] Oct 09 10:32:14 crc kubenswrapper[4923]: I1009 10:32:14.626861 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bede9ca-eff0-422f-bc53-b156346d993f" path="/var/lib/kubelet/pods/3bede9ca-eff0-422f-bc53-b156346d993f/volumes" Oct 09 10:32:14 crc kubenswrapper[4923]: I1009 10:32:14.627995 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee194426-1c9c-49f6-b112-8bf243b2d8ca" path="/var/lib/kubelet/pods/ee194426-1c9c-49f6-b112-8bf243b2d8ca/volumes" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.004899 4923 scope.go:117] "RemoveContainer" containerID="d4e767d072551c1fdb7bb73cdd997ed2cdb0489605c1f682ea5f0080487fbe77" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.057844 4923 scope.go:117] "RemoveContainer" containerID="bd71ad4b8fbe98ab192d364066620e857c91a7e87102664e6412ec7329fa015d" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.095060 4923 scope.go:117] "RemoveContainer" containerID="44e68a14bb6ea8b8055bb92d53e0cbc3689a8a32507e8bb1adef365e021e5223" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.151852 4923 scope.go:117] "RemoveContainer" containerID="9c70a17e35e220ec215f760024df38d7563a480010a73842a8a9e2b8ad6ea8c7" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.191924 4923 scope.go:117] "RemoveContainer" containerID="01b7f8b396fe8294d9d211391daeade41d79f6378e88876e0122b461e2b48c22" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.236615 4923 scope.go:117] "RemoveContainer" containerID="52dfbd33a4b04e87362c2cc041044c5e407650b00fe639046e1159b961498545" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.286496 4923 scope.go:117] "RemoveContainer" containerID="4f9e9f9df6fd102dee0a5835a60d51311b8114f07ac5afb39e19889b26c73ab7" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.309320 4923 scope.go:117] "RemoveContainer" containerID="df7c565a2ba98b9c9a71d71dcec291fecef54730336abc5d1f3be7d4b2eb3fa0" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.328573 4923 scope.go:117] "RemoveContainer" containerID="12a8551a0ecabb2686d2b0b788853b7b3fe6e2d501c97bc1bb7ff5b30b3cf0b7" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.350964 4923 scope.go:117] "RemoveContainer" containerID="cc98537e28beea7ef9b746322c1eeaf43f9ecbaf87fc70deefa5fba557733d0e" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.377848 4923 scope.go:117] "RemoveContainer" containerID="650c63ee4e7e5f6c3a703abd4efeec520b335c2ebd3c443573dac709d2abab09" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.400870 4923 scope.go:117] "RemoveContainer" containerID="2d366a97b0924e8d3eeb3478d7fc9dfeec921a6c51bda9d8b02cbb7e579c03a4" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.436388 4923 scope.go:117] "RemoveContainer" containerID="98e77b5646898ba123a7413d3eaaffc4bb39735dc78eb2a2ddeab9b565f39be8" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.490126 4923 scope.go:117] "RemoveContainer" containerID="61c0e03f87d6e78e0d9732565997c372418a8a985104a15d889346157ca27e70" Oct 09 10:32:17 crc kubenswrapper[4923]: I1009 10:32:17.512274 4923 scope.go:117] "RemoveContainer" containerID="dfc10e4bff23b6038dbd423ffc673adfbd68742eb983ad8f2ee294ce53c39893" Oct 09 10:32:22 crc kubenswrapper[4923]: I1009 10:32:22.037507 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wbnxf"] Oct 09 10:32:22 crc kubenswrapper[4923]: I1009 10:32:22.048404 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wbnxf"] Oct 09 10:32:22 crc kubenswrapper[4923]: I1009 10:32:22.602316 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:32:22 crc kubenswrapper[4923]: E1009 10:32:22.603266 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:32:22 crc kubenswrapper[4923]: I1009 10:32:22.614435 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39c4d2f2-2bdd-417d-ab59-a99054774436" path="/var/lib/kubelet/pods/39c4d2f2-2bdd-417d-ab59-a99054774436/volumes" Oct 09 10:32:34 crc kubenswrapper[4923]: I1009 10:32:34.610585 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:32:34 crc kubenswrapper[4923]: E1009 10:32:34.612501 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.374306 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tshjz"] Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.377831 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.391689 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tshjz"] Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.491385 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvds7\" (UniqueName: \"kubernetes.io/projected/9a160c9e-8705-4fde-8c20-a398d64ce8de-kube-api-access-bvds7\") pod \"certified-operators-tshjz\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.491964 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-catalog-content\") pod \"certified-operators-tshjz\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.492037 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-utilities\") pod \"certified-operators-tshjz\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.594716 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvds7\" (UniqueName: \"kubernetes.io/projected/9a160c9e-8705-4fde-8c20-a398d64ce8de-kube-api-access-bvds7\") pod \"certified-operators-tshjz\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.594841 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-catalog-content\") pod \"certified-operators-tshjz\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.594892 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-utilities\") pod \"certified-operators-tshjz\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.595592 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-catalog-content\") pod \"certified-operators-tshjz\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.595628 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-utilities\") pod \"certified-operators-tshjz\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.616875 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvds7\" (UniqueName: \"kubernetes.io/projected/9a160c9e-8705-4fde-8c20-a398d64ce8de-kube-api-access-bvds7\") pod \"certified-operators-tshjz\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:38 crc kubenswrapper[4923]: I1009 10:32:38.714694 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:39 crc kubenswrapper[4923]: I1009 10:32:39.240879 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tshjz"] Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.218805 4923 generic.go:334] "Generic (PLEG): container finished" podID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerID="023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4" exitCode=0 Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.218991 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshjz" event={"ID":"9a160c9e-8705-4fde-8c20-a398d64ce8de","Type":"ContainerDied","Data":"023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4"} Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.219338 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshjz" event={"ID":"9a160c9e-8705-4fde-8c20-a398d64ce8de","Type":"ContainerStarted","Data":"1cee11c1b2d6713e2857c710c70b50199d49697dcce86dc4c8bb8996b38e745c"} Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.758456 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t44wx"] Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.761281 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.775935 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t44wx"] Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.853614 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxb5j\" (UniqueName: \"kubernetes.io/projected/fe74d812-8294-4885-bb17-bddd0e7af92d-kube-api-access-nxb5j\") pod \"redhat-operators-t44wx\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.853768 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-catalog-content\") pod \"redhat-operators-t44wx\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.853843 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-utilities\") pod \"redhat-operators-t44wx\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.958416 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-catalog-content\") pod \"redhat-operators-t44wx\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.958494 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-utilities\") pod \"redhat-operators-t44wx\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.958598 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxb5j\" (UniqueName: \"kubernetes.io/projected/fe74d812-8294-4885-bb17-bddd0e7af92d-kube-api-access-nxb5j\") pod \"redhat-operators-t44wx\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.959070 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-catalog-content\") pod \"redhat-operators-t44wx\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.959177 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-utilities\") pod \"redhat-operators-t44wx\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:40 crc kubenswrapper[4923]: I1009 10:32:40.984202 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxb5j\" (UniqueName: \"kubernetes.io/projected/fe74d812-8294-4885-bb17-bddd0e7af92d-kube-api-access-nxb5j\") pod \"redhat-operators-t44wx\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:41 crc kubenswrapper[4923]: I1009 10:32:41.083593 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:41 crc kubenswrapper[4923]: I1009 10:32:41.580160 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t44wx"] Oct 09 10:32:42 crc kubenswrapper[4923]: I1009 10:32:42.251528 4923 generic.go:334] "Generic (PLEG): container finished" podID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerID="1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547" exitCode=0 Oct 09 10:32:42 crc kubenswrapper[4923]: I1009 10:32:42.251597 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshjz" event={"ID":"9a160c9e-8705-4fde-8c20-a398d64ce8de","Type":"ContainerDied","Data":"1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547"} Oct 09 10:32:42 crc kubenswrapper[4923]: I1009 10:32:42.254614 4923 generic.go:334] "Generic (PLEG): container finished" podID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerID="266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65" exitCode=0 Oct 09 10:32:42 crc kubenswrapper[4923]: I1009 10:32:42.254646 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t44wx" event={"ID":"fe74d812-8294-4885-bb17-bddd0e7af92d","Type":"ContainerDied","Data":"266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65"} Oct 09 10:32:42 crc kubenswrapper[4923]: I1009 10:32:42.254674 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t44wx" event={"ID":"fe74d812-8294-4885-bb17-bddd0e7af92d","Type":"ContainerStarted","Data":"7538e815c8c4121b2d1e149c95e314fb0f9faa98de3c8ab7660a53367821df39"} Oct 09 10:32:43 crc kubenswrapper[4923]: I1009 10:32:43.269675 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshjz" event={"ID":"9a160c9e-8705-4fde-8c20-a398d64ce8de","Type":"ContainerStarted","Data":"db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49"} Oct 09 10:32:43 crc kubenswrapper[4923]: I1009 10:32:43.298975 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tshjz" podStartSLOduration=2.6742656350000003 podStartE2EDuration="5.298943397s" podCreationTimestamp="2025-10-09 10:32:38 +0000 UTC" firstStartedPulling="2025-10-09 10:32:40.221276531 +0000 UTC m=+1646.289458287" lastFinishedPulling="2025-10-09 10:32:42.845954293 +0000 UTC m=+1648.914136049" observedRunningTime="2025-10-09 10:32:43.291678544 +0000 UTC m=+1649.359860300" watchObservedRunningTime="2025-10-09 10:32:43.298943397 +0000 UTC m=+1649.367125143" Oct 09 10:32:44 crc kubenswrapper[4923]: I1009 10:32:44.284245 4923 generic.go:334] "Generic (PLEG): container finished" podID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerID="b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1" exitCode=0 Oct 09 10:32:44 crc kubenswrapper[4923]: I1009 10:32:44.284355 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t44wx" event={"ID":"fe74d812-8294-4885-bb17-bddd0e7af92d","Type":"ContainerDied","Data":"b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1"} Oct 09 10:32:47 crc kubenswrapper[4923]: I1009 10:32:47.340874 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t44wx" event={"ID":"fe74d812-8294-4885-bb17-bddd0e7af92d","Type":"ContainerStarted","Data":"91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33"} Oct 09 10:32:47 crc kubenswrapper[4923]: I1009 10:32:47.367470 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t44wx" podStartSLOduration=3.493109199 podStartE2EDuration="7.367443259s" podCreationTimestamp="2025-10-09 10:32:40 +0000 UTC" firstStartedPulling="2025-10-09 10:32:42.257030469 +0000 UTC m=+1648.325212225" lastFinishedPulling="2025-10-09 10:32:46.131364529 +0000 UTC m=+1652.199546285" observedRunningTime="2025-10-09 10:32:47.364216839 +0000 UTC m=+1653.432398595" watchObservedRunningTime="2025-10-09 10:32:47.367443259 +0000 UTC m=+1653.435625015" Oct 09 10:32:48 crc kubenswrapper[4923]: I1009 10:32:48.715491 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:48 crc kubenswrapper[4923]: I1009 10:32:48.716203 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:48 crc kubenswrapper[4923]: I1009 10:32:48.784035 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:49 crc kubenswrapper[4923]: I1009 10:32:49.435431 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:49 crc kubenswrapper[4923]: I1009 10:32:49.602219 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:32:49 crc kubenswrapper[4923]: E1009 10:32:49.602609 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:32:49 crc kubenswrapper[4923]: I1009 10:32:49.740366 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tshjz"] Oct 09 10:32:51 crc kubenswrapper[4923]: I1009 10:32:51.084456 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:51 crc kubenswrapper[4923]: I1009 10:32:51.084701 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:51 crc kubenswrapper[4923]: I1009 10:32:51.140572 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:51 crc kubenswrapper[4923]: I1009 10:32:51.405077 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tshjz" podUID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerName="registry-server" containerID="cri-o://db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49" gracePeriod=2 Oct 09 10:32:51 crc kubenswrapper[4923]: I1009 10:32:51.465249 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:51 crc kubenswrapper[4923]: I1009 10:32:51.890004 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.093923 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-utilities\") pod \"9a160c9e-8705-4fde-8c20-a398d64ce8de\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.094133 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-catalog-content\") pod \"9a160c9e-8705-4fde-8c20-a398d64ce8de\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.094307 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvds7\" (UniqueName: \"kubernetes.io/projected/9a160c9e-8705-4fde-8c20-a398d64ce8de-kube-api-access-bvds7\") pod \"9a160c9e-8705-4fde-8c20-a398d64ce8de\" (UID: \"9a160c9e-8705-4fde-8c20-a398d64ce8de\") " Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.094815 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-utilities" (OuterVolumeSpecName: "utilities") pod "9a160c9e-8705-4fde-8c20-a398d64ce8de" (UID: "9a160c9e-8705-4fde-8c20-a398d64ce8de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.104465 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a160c9e-8705-4fde-8c20-a398d64ce8de-kube-api-access-bvds7" (OuterVolumeSpecName: "kube-api-access-bvds7") pod "9a160c9e-8705-4fde-8c20-a398d64ce8de" (UID: "9a160c9e-8705-4fde-8c20-a398d64ce8de"). InnerVolumeSpecName "kube-api-access-bvds7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.151087 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a160c9e-8705-4fde-8c20-a398d64ce8de" (UID: "9a160c9e-8705-4fde-8c20-a398d64ce8de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.196798 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvds7\" (UniqueName: \"kubernetes.io/projected/9a160c9e-8705-4fde-8c20-a398d64ce8de-kube-api-access-bvds7\") on node \"crc\" DevicePath \"\"" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.196856 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.196868 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a160c9e-8705-4fde-8c20-a398d64ce8de-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.343192 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t44wx"] Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.416956 4923 generic.go:334] "Generic (PLEG): container finished" podID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerID="db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49" exitCode=0 Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.417035 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshjz" event={"ID":"9a160c9e-8705-4fde-8c20-a398d64ce8de","Type":"ContainerDied","Data":"db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49"} Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.417086 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tshjz" event={"ID":"9a160c9e-8705-4fde-8c20-a398d64ce8de","Type":"ContainerDied","Data":"1cee11c1b2d6713e2857c710c70b50199d49697dcce86dc4c8bb8996b38e745c"} Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.417112 4923 scope.go:117] "RemoveContainer" containerID="db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.417048 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tshjz" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.440450 4923 scope.go:117] "RemoveContainer" containerID="1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.470084 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tshjz"] Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.479252 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tshjz"] Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.490233 4923 scope.go:117] "RemoveContainer" containerID="023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.514264 4923 scope.go:117] "RemoveContainer" containerID="db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49" Oct 09 10:32:52 crc kubenswrapper[4923]: E1009 10:32:52.514738 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49\": container with ID starting with db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49 not found: ID does not exist" containerID="db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.514828 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49"} err="failed to get container status \"db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49\": rpc error: code = NotFound desc = could not find container \"db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49\": container with ID starting with db3e1b9b75434ad1505f5c98ffa7c4b305c3b93e7f1599b7609cd30c7192da49 not found: ID does not exist" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.514865 4923 scope.go:117] "RemoveContainer" containerID="1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547" Oct 09 10:32:52 crc kubenswrapper[4923]: E1009 10:32:52.516331 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547\": container with ID starting with 1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547 not found: ID does not exist" containerID="1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.516395 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547"} err="failed to get container status \"1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547\": rpc error: code = NotFound desc = could not find container \"1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547\": container with ID starting with 1f0e7b9104a14e4a9cd9cc39df5e4a1de166572f0fbd96d488e7425d4c78c547 not found: ID does not exist" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.516435 4923 scope.go:117] "RemoveContainer" containerID="023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4" Oct 09 10:32:52 crc kubenswrapper[4923]: E1009 10:32:52.516825 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4\": container with ID starting with 023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4 not found: ID does not exist" containerID="023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.516858 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4"} err="failed to get container status \"023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4\": rpc error: code = NotFound desc = could not find container \"023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4\": container with ID starting with 023706e5117633cbc2c26cbf85ba190179fab0119a839731ffcd7a72d4c9daa4 not found: ID does not exist" Oct 09 10:32:52 crc kubenswrapper[4923]: I1009 10:32:52.624252 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a160c9e-8705-4fde-8c20-a398d64ce8de" path="/var/lib/kubelet/pods/9a160c9e-8705-4fde-8c20-a398d64ce8de/volumes" Oct 09 10:32:53 crc kubenswrapper[4923]: I1009 10:32:53.428816 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-t44wx" podUID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerName="registry-server" containerID="cri-o://91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33" gracePeriod=2 Oct 09 10:32:53 crc kubenswrapper[4923]: I1009 10:32:53.901334 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.039444 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxb5j\" (UniqueName: \"kubernetes.io/projected/fe74d812-8294-4885-bb17-bddd0e7af92d-kube-api-access-nxb5j\") pod \"fe74d812-8294-4885-bb17-bddd0e7af92d\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.039553 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-utilities\") pod \"fe74d812-8294-4885-bb17-bddd0e7af92d\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.039775 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-catalog-content\") pod \"fe74d812-8294-4885-bb17-bddd0e7af92d\" (UID: \"fe74d812-8294-4885-bb17-bddd0e7af92d\") " Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.041190 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-utilities" (OuterVolumeSpecName: "utilities") pod "fe74d812-8294-4885-bb17-bddd0e7af92d" (UID: "fe74d812-8294-4885-bb17-bddd0e7af92d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.048191 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe74d812-8294-4885-bb17-bddd0e7af92d-kube-api-access-nxb5j" (OuterVolumeSpecName: "kube-api-access-nxb5j") pod "fe74d812-8294-4885-bb17-bddd0e7af92d" (UID: "fe74d812-8294-4885-bb17-bddd0e7af92d"). InnerVolumeSpecName "kube-api-access-nxb5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.057149 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-5jlhb"] Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.069345 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-5jlhb"] Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.143473 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxb5j\" (UniqueName: \"kubernetes.io/projected/fe74d812-8294-4885-bb17-bddd0e7af92d-kube-api-access-nxb5j\") on node \"crc\" DevicePath \"\"" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.143545 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.152806 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe74d812-8294-4885-bb17-bddd0e7af92d" (UID: "fe74d812-8294-4885-bb17-bddd0e7af92d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.245636 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe74d812-8294-4885-bb17-bddd0e7af92d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.448937 4923 generic.go:334] "Generic (PLEG): container finished" podID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerID="91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33" exitCode=0 Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.449701 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t44wx" event={"ID":"fe74d812-8294-4885-bb17-bddd0e7af92d","Type":"ContainerDied","Data":"91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33"} Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.449814 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t44wx" event={"ID":"fe74d812-8294-4885-bb17-bddd0e7af92d","Type":"ContainerDied","Data":"7538e815c8c4121b2d1e149c95e314fb0f9faa98de3c8ab7660a53367821df39"} Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.449885 4923 scope.go:117] "RemoveContainer" containerID="91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.450088 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t44wx" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.498883 4923 scope.go:117] "RemoveContainer" containerID="b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.501511 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-t44wx"] Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.510491 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-t44wx"] Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.526188 4923 scope.go:117] "RemoveContainer" containerID="266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.564967 4923 scope.go:117] "RemoveContainer" containerID="91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33" Oct 09 10:32:54 crc kubenswrapper[4923]: E1009 10:32:54.565499 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33\": container with ID starting with 91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33 not found: ID does not exist" containerID="91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.565593 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33"} err="failed to get container status \"91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33\": rpc error: code = NotFound desc = could not find container \"91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33\": container with ID starting with 91ceadea6c8fec04a1fda06d3f0ff15d1b9639dd2dddf0cc15f8aaa615f8bd33 not found: ID does not exist" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.565675 4923 scope.go:117] "RemoveContainer" containerID="b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1" Oct 09 10:32:54 crc kubenswrapper[4923]: E1009 10:32:54.566266 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1\": container with ID starting with b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1 not found: ID does not exist" containerID="b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.566334 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1"} err="failed to get container status \"b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1\": rpc error: code = NotFound desc = could not find container \"b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1\": container with ID starting with b724b3be4cf3ee92e47cb57652a23e16da84bef8715dc3ce8c29eaaf312bfca1 not found: ID does not exist" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.566393 4923 scope.go:117] "RemoveContainer" containerID="266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65" Oct 09 10:32:54 crc kubenswrapper[4923]: E1009 10:32:54.567005 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65\": container with ID starting with 266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65 not found: ID does not exist" containerID="266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.567039 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65"} err="failed to get container status \"266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65\": rpc error: code = NotFound desc = could not find container \"266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65\": container with ID starting with 266fd79027e5c68ca5c340fba3483b4fff5204ff94dac9c1bb766006a737ce65 not found: ID does not exist" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.612353 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03ad58cb-353a-493e-a7d0-b5c66c3261bf" path="/var/lib/kubelet/pods/03ad58cb-353a-493e-a7d0-b5c66c3261bf/volumes" Oct 09 10:32:54 crc kubenswrapper[4923]: I1009 10:32:54.613431 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe74d812-8294-4885-bb17-bddd0e7af92d" path="/var/lib/kubelet/pods/fe74d812-8294-4885-bb17-bddd0e7af92d/volumes" Oct 09 10:33:00 crc kubenswrapper[4923]: I1009 10:33:00.601718 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:33:00 crc kubenswrapper[4923]: E1009 10:33:00.602894 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:33:12 crc kubenswrapper[4923]: I1009 10:33:12.603016 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:33:12 crc kubenswrapper[4923]: E1009 10:33:12.606050 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:33:17 crc kubenswrapper[4923]: I1009 10:33:17.054779 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2m7bb"] Oct 09 10:33:17 crc kubenswrapper[4923]: I1009 10:33:17.067772 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-665vn"] Oct 09 10:33:17 crc kubenswrapper[4923]: I1009 10:33:17.077005 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2m7bb"] Oct 09 10:33:17 crc kubenswrapper[4923]: I1009 10:33:17.085150 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-665vn"] Oct 09 10:33:17 crc kubenswrapper[4923]: I1009 10:33:17.773493 4923 scope.go:117] "RemoveContainer" containerID="d8f0aa1852932b3fb1536d7b44c02302c13c8c4791f9719aca6fd74693698c44" Oct 09 10:33:17 crc kubenswrapper[4923]: I1009 10:33:17.851123 4923 scope.go:117] "RemoveContainer" containerID="be8f5935fa858eb0b33b9f22b84de796c7dab680e5ce7948cf4065b5202b3c16" Oct 09 10:33:18 crc kubenswrapper[4923]: I1009 10:33:18.032041 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-xq58z"] Oct 09 10:33:18 crc kubenswrapper[4923]: I1009 10:33:18.040859 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-xq58z"] Oct 09 10:33:18 crc kubenswrapper[4923]: I1009 10:33:18.617071 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31f10461-abd0-4c4d-bdb5-3da1f1e6ed07" path="/var/lib/kubelet/pods/31f10461-abd0-4c4d-bdb5-3da1f1e6ed07/volumes" Oct 09 10:33:18 crc kubenswrapper[4923]: I1009 10:33:18.618045 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85c95222-efd5-4bd6-ab01-c51eb21abb4c" path="/var/lib/kubelet/pods/85c95222-efd5-4bd6-ab01-c51eb21abb4c/volumes" Oct 09 10:33:18 crc kubenswrapper[4923]: I1009 10:33:18.618847 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f250b2a7-fb69-4c40-b186-238f74472a44" path="/var/lib/kubelet/pods/f250b2a7-fb69-4c40-b186-238f74472a44/volumes" Oct 09 10:33:19 crc kubenswrapper[4923]: I1009 10:33:19.696000 4923 generic.go:334] "Generic (PLEG): container finished" podID="00251bc8-62c6-4f4e-8aa0-f8da07570f3a" containerID="6f0cd671f826c9d29056422776a0cf628ae70499aefc6b58dccc8b71acb825ea" exitCode=0 Oct 09 10:33:19 crc kubenswrapper[4923]: I1009 10:33:19.696086 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" event={"ID":"00251bc8-62c6-4f4e-8aa0-f8da07570f3a","Type":"ContainerDied","Data":"6f0cd671f826c9d29056422776a0cf628ae70499aefc6b58dccc8b71acb825ea"} Oct 09 10:33:20 crc kubenswrapper[4923]: I1009 10:33:20.035180 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-5ll8s"] Oct 09 10:33:20 crc kubenswrapper[4923]: I1009 10:33:20.042790 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-5ll8s"] Oct 09 10:33:20 crc kubenswrapper[4923]: I1009 10:33:20.613805 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd" path="/var/lib/kubelet/pods/0c688fdd-7d04-4a0a-9074-94c9f1b8a5fd/volumes" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.169930 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.290862 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-inventory\") pod \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.290987 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-ssh-key\") pod \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.291244 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwrgt\" (UniqueName: \"kubernetes.io/projected/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-kube-api-access-cwrgt\") pod \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\" (UID: \"00251bc8-62c6-4f4e-8aa0-f8da07570f3a\") " Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.300192 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-kube-api-access-cwrgt" (OuterVolumeSpecName: "kube-api-access-cwrgt") pod "00251bc8-62c6-4f4e-8aa0-f8da07570f3a" (UID: "00251bc8-62c6-4f4e-8aa0-f8da07570f3a"). InnerVolumeSpecName "kube-api-access-cwrgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.327552 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "00251bc8-62c6-4f4e-8aa0-f8da07570f3a" (UID: "00251bc8-62c6-4f4e-8aa0-f8da07570f3a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.327695 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-inventory" (OuterVolumeSpecName: "inventory") pod "00251bc8-62c6-4f4e-8aa0-f8da07570f3a" (UID: "00251bc8-62c6-4f4e-8aa0-f8da07570f3a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.395224 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.395276 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.395290 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwrgt\" (UniqueName: \"kubernetes.io/projected/00251bc8-62c6-4f4e-8aa0-f8da07570f3a-kube-api-access-cwrgt\") on node \"crc\" DevicePath \"\"" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.717460 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" event={"ID":"00251bc8-62c6-4f4e-8aa0-f8da07570f3a","Type":"ContainerDied","Data":"f008a5c2a03529e245025565ddfe6ff35b0945349b0d044dd926bffbaf5cefa9"} Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.717537 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f008a5c2a03529e245025565ddfe6ff35b0945349b0d044dd926bffbaf5cefa9" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.717506 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.797570 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8"] Oct 09 10:33:21 crc kubenswrapper[4923]: E1009 10:33:21.798385 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerName="extract-utilities" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798415 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerName="extract-utilities" Oct 09 10:33:21 crc kubenswrapper[4923]: E1009 10:33:21.798449 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerName="registry-server" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798459 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerName="registry-server" Oct 09 10:33:21 crc kubenswrapper[4923]: E1009 10:33:21.798468 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00251bc8-62c6-4f4e-8aa0-f8da07570f3a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798477 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="00251bc8-62c6-4f4e-8aa0-f8da07570f3a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:33:21 crc kubenswrapper[4923]: E1009 10:33:21.798488 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerName="extract-utilities" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798494 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerName="extract-utilities" Oct 09 10:33:21 crc kubenswrapper[4923]: E1009 10:33:21.798502 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerName="registry-server" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798508 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerName="registry-server" Oct 09 10:33:21 crc kubenswrapper[4923]: E1009 10:33:21.798526 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerName="extract-content" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798532 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerName="extract-content" Oct 09 10:33:21 crc kubenswrapper[4923]: E1009 10:33:21.798552 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerName="extract-content" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798558 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerName="extract-content" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798725 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="00251bc8-62c6-4f4e-8aa0-f8da07570f3a" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798753 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a160c9e-8705-4fde-8c20-a398d64ce8de" containerName="registry-server" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.798779 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe74d812-8294-4885-bb17-bddd0e7af92d" containerName="registry-server" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.799576 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.802334 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.802686 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.803236 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.807067 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.811847 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8"] Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.907516 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-njnq8\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.907624 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5g2h4\" (UniqueName: \"kubernetes.io/projected/ff2e838f-c4ce-43db-bb07-4cd0191d0079-kube-api-access-5g2h4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-njnq8\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:21 crc kubenswrapper[4923]: I1009 10:33:21.907915 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-njnq8\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:22 crc kubenswrapper[4923]: I1009 10:33:22.010346 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-njnq8\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:22 crc kubenswrapper[4923]: I1009 10:33:22.010486 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5g2h4\" (UniqueName: \"kubernetes.io/projected/ff2e838f-c4ce-43db-bb07-4cd0191d0079-kube-api-access-5g2h4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-njnq8\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:22 crc kubenswrapper[4923]: I1009 10:33:22.010528 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-njnq8\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:22 crc kubenswrapper[4923]: I1009 10:33:22.014753 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-njnq8\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:22 crc kubenswrapper[4923]: I1009 10:33:22.031525 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-njnq8\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:22 crc kubenswrapper[4923]: I1009 10:33:22.033244 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5g2h4\" (UniqueName: \"kubernetes.io/projected/ff2e838f-c4ce-43db-bb07-4cd0191d0079-kube-api-access-5g2h4\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-njnq8\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:22 crc kubenswrapper[4923]: I1009 10:33:22.122500 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:22 crc kubenswrapper[4923]: I1009 10:33:22.709147 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8"] Oct 09 10:33:22 crc kubenswrapper[4923]: W1009 10:33:22.719291 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff2e838f_c4ce_43db_bb07_4cd0191d0079.slice/crio-953a06cfac6b7a49a88df422a21a20aa36377fcdb2f4109d66876700076f6733 WatchSource:0}: Error finding container 953a06cfac6b7a49a88df422a21a20aa36377fcdb2f4109d66876700076f6733: Status 404 returned error can't find the container with id 953a06cfac6b7a49a88df422a21a20aa36377fcdb2f4109d66876700076f6733 Oct 09 10:33:23 crc kubenswrapper[4923]: I1009 10:33:23.765281 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" event={"ID":"ff2e838f-c4ce-43db-bb07-4cd0191d0079","Type":"ContainerStarted","Data":"c6e06ea5c1a11c5a00add805c5240fdd44f274e5f7bd9baf666ba0b71e725e67"} Oct 09 10:33:23 crc kubenswrapper[4923]: I1009 10:33:23.767773 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" event={"ID":"ff2e838f-c4ce-43db-bb07-4cd0191d0079","Type":"ContainerStarted","Data":"953a06cfac6b7a49a88df422a21a20aa36377fcdb2f4109d66876700076f6733"} Oct 09 10:33:23 crc kubenswrapper[4923]: I1009 10:33:23.799053 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" podStartSLOduration=2.191353532 podStartE2EDuration="2.799025858s" podCreationTimestamp="2025-10-09 10:33:21 +0000 UTC" firstStartedPulling="2025-10-09 10:33:22.727023593 +0000 UTC m=+1688.795205369" lastFinishedPulling="2025-10-09 10:33:23.334695919 +0000 UTC m=+1689.402877695" observedRunningTime="2025-10-09 10:33:23.787151458 +0000 UTC m=+1689.855333224" watchObservedRunningTime="2025-10-09 10:33:23.799025858 +0000 UTC m=+1689.867207614" Oct 09 10:33:26 crc kubenswrapper[4923]: I1009 10:33:26.602099 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:33:26 crc kubenswrapper[4923]: E1009 10:33:26.602977 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:33:28 crc kubenswrapper[4923]: I1009 10:33:28.815537 4923 generic.go:334] "Generic (PLEG): container finished" podID="ff2e838f-c4ce-43db-bb07-4cd0191d0079" containerID="c6e06ea5c1a11c5a00add805c5240fdd44f274e5f7bd9baf666ba0b71e725e67" exitCode=0 Oct 09 10:33:28 crc kubenswrapper[4923]: I1009 10:33:28.815598 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" event={"ID":"ff2e838f-c4ce-43db-bb07-4cd0191d0079","Type":"ContainerDied","Data":"c6e06ea5c1a11c5a00add805c5240fdd44f274e5f7bd9baf666ba0b71e725e67"} Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.308325 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.499982 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-inventory\") pod \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.500123 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5g2h4\" (UniqueName: \"kubernetes.io/projected/ff2e838f-c4ce-43db-bb07-4cd0191d0079-kube-api-access-5g2h4\") pod \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.500244 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-ssh-key\") pod \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\" (UID: \"ff2e838f-c4ce-43db-bb07-4cd0191d0079\") " Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.507501 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff2e838f-c4ce-43db-bb07-4cd0191d0079-kube-api-access-5g2h4" (OuterVolumeSpecName: "kube-api-access-5g2h4") pod "ff2e838f-c4ce-43db-bb07-4cd0191d0079" (UID: "ff2e838f-c4ce-43db-bb07-4cd0191d0079"). InnerVolumeSpecName "kube-api-access-5g2h4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.530559 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ff2e838f-c4ce-43db-bb07-4cd0191d0079" (UID: "ff2e838f-c4ce-43db-bb07-4cd0191d0079"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.534264 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-inventory" (OuterVolumeSpecName: "inventory") pod "ff2e838f-c4ce-43db-bb07-4cd0191d0079" (UID: "ff2e838f-c4ce-43db-bb07-4cd0191d0079"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.602379 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.602423 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ff2e838f-c4ce-43db-bb07-4cd0191d0079-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.602437 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5g2h4\" (UniqueName: \"kubernetes.io/projected/ff2e838f-c4ce-43db-bb07-4cd0191d0079-kube-api-access-5g2h4\") on node \"crc\" DevicePath \"\"" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.839570 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" event={"ID":"ff2e838f-c4ce-43db-bb07-4cd0191d0079","Type":"ContainerDied","Data":"953a06cfac6b7a49a88df422a21a20aa36377fcdb2f4109d66876700076f6733"} Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.839633 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="953a06cfac6b7a49a88df422a21a20aa36377fcdb2f4109d66876700076f6733" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.839953 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.941396 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb"] Oct 09 10:33:30 crc kubenswrapper[4923]: E1009 10:33:30.942032 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff2e838f-c4ce-43db-bb07-4cd0191d0079" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.942073 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff2e838f-c4ce-43db-bb07-4cd0191d0079" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.942444 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff2e838f-c4ce-43db-bb07-4cd0191d0079" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.943355 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.946198 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.946525 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.946807 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.946865 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:33:30 crc kubenswrapper[4923]: I1009 10:33:30.967181 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb"] Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.111857 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfcpb\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.111975 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v94c2\" (UniqueName: \"kubernetes.io/projected/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-kube-api-access-v94c2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfcpb\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.112069 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfcpb\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.215262 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfcpb\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.215569 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfcpb\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.215669 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v94c2\" (UniqueName: \"kubernetes.io/projected/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-kube-api-access-v94c2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfcpb\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.239426 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfcpb\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.239434 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfcpb\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.246028 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v94c2\" (UniqueName: \"kubernetes.io/projected/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-kube-api-access-v94c2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-sfcpb\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.264774 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.826129 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb"] Oct 09 10:33:31 crc kubenswrapper[4923]: I1009 10:33:31.853912 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" event={"ID":"0db9879d-8e0a-4d29-9168-caf8e16d5c3d","Type":"ContainerStarted","Data":"b72f22392fc3adc615a4d5963b1705558ae485cd9f601d67db9bc14090b47b38"} Oct 09 10:33:33 crc kubenswrapper[4923]: I1009 10:33:33.884864 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" event={"ID":"0db9879d-8e0a-4d29-9168-caf8e16d5c3d","Type":"ContainerStarted","Data":"6431403dfee5423036e307a91cd66a480d43d6359485ccedc4cba1d163f7b4cc"} Oct 09 10:33:33 crc kubenswrapper[4923]: I1009 10:33:33.905366 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" podStartSLOduration=3.136725448 podStartE2EDuration="3.905338793s" podCreationTimestamp="2025-10-09 10:33:30 +0000 UTC" firstStartedPulling="2025-10-09 10:33:31.833987304 +0000 UTC m=+1697.902169060" lastFinishedPulling="2025-10-09 10:33:32.602600639 +0000 UTC m=+1698.670782405" observedRunningTime="2025-10-09 10:33:33.904055457 +0000 UTC m=+1699.972237223" watchObservedRunningTime="2025-10-09 10:33:33.905338793 +0000 UTC m=+1699.973520549" Oct 09 10:33:38 crc kubenswrapper[4923]: I1009 10:33:38.601276 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:33:38 crc kubenswrapper[4923]: E1009 10:33:38.602076 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:33:49 crc kubenswrapper[4923]: I1009 10:33:49.602448 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:33:49 crc kubenswrapper[4923]: E1009 10:33:49.604906 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:34:02 crc kubenswrapper[4923]: I1009 10:34:02.602387 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:34:02 crc kubenswrapper[4923]: E1009 10:34:02.603436 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:34:09 crc kubenswrapper[4923]: I1009 10:34:09.063912 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-ckntf"] Oct 09 10:34:09 crc kubenswrapper[4923]: I1009 10:34:09.079386 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-nwwwr"] Oct 09 10:34:09 crc kubenswrapper[4923]: I1009 10:34:09.089794 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-7pgnz"] Oct 09 10:34:09 crc kubenswrapper[4923]: I1009 10:34:09.101085 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-ckntf"] Oct 09 10:34:09 crc kubenswrapper[4923]: I1009 10:34:09.111370 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-nwwwr"] Oct 09 10:34:09 crc kubenswrapper[4923]: I1009 10:34:09.119572 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-7pgnz"] Oct 09 10:34:10 crc kubenswrapper[4923]: I1009 10:34:10.615644 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4747efe7-6523-4c92-82d3-ce5d2a8b2868" path="/var/lib/kubelet/pods/4747efe7-6523-4c92-82d3-ce5d2a8b2868/volumes" Oct 09 10:34:10 crc kubenswrapper[4923]: I1009 10:34:10.616884 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="667c88d0-d5c3-416c-b166-b68f127ce657" path="/var/lib/kubelet/pods/667c88d0-d5c3-416c-b166-b68f127ce657/volumes" Oct 09 10:34:10 crc kubenswrapper[4923]: I1009 10:34:10.617415 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a32a19c5-4480-4d2b-9563-208e6a0ced35" path="/var/lib/kubelet/pods/a32a19c5-4480-4d2b-9563-208e6a0ced35/volumes" Oct 09 10:34:11 crc kubenswrapper[4923]: I1009 10:34:11.287289 4923 generic.go:334] "Generic (PLEG): container finished" podID="0db9879d-8e0a-4d29-9168-caf8e16d5c3d" containerID="6431403dfee5423036e307a91cd66a480d43d6359485ccedc4cba1d163f7b4cc" exitCode=0 Oct 09 10:34:11 crc kubenswrapper[4923]: I1009 10:34:11.287343 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" event={"ID":"0db9879d-8e0a-4d29-9168-caf8e16d5c3d","Type":"ContainerDied","Data":"6431403dfee5423036e307a91cd66a480d43d6359485ccedc4cba1d163f7b4cc"} Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.728084 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.860006 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v94c2\" (UniqueName: \"kubernetes.io/projected/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-kube-api-access-v94c2\") pod \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.860161 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-ssh-key\") pod \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.860264 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-inventory\") pod \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\" (UID: \"0db9879d-8e0a-4d29-9168-caf8e16d5c3d\") " Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.868868 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-kube-api-access-v94c2" (OuterVolumeSpecName: "kube-api-access-v94c2") pod "0db9879d-8e0a-4d29-9168-caf8e16d5c3d" (UID: "0db9879d-8e0a-4d29-9168-caf8e16d5c3d"). InnerVolumeSpecName "kube-api-access-v94c2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.902694 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0db9879d-8e0a-4d29-9168-caf8e16d5c3d" (UID: "0db9879d-8e0a-4d29-9168-caf8e16d5c3d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.915845 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-inventory" (OuterVolumeSpecName: "inventory") pod "0db9879d-8e0a-4d29-9168-caf8e16d5c3d" (UID: "0db9879d-8e0a-4d29-9168-caf8e16d5c3d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.962816 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v94c2\" (UniqueName: \"kubernetes.io/projected/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-kube-api-access-v94c2\") on node \"crc\" DevicePath \"\"" Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.962872 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:34:12 crc kubenswrapper[4923]: I1009 10:34:12.962888 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0db9879d-8e0a-4d29-9168-caf8e16d5c3d-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.356671 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" event={"ID":"0db9879d-8e0a-4d29-9168-caf8e16d5c3d","Type":"ContainerDied","Data":"b72f22392fc3adc615a4d5963b1705558ae485cd9f601d67db9bc14090b47b38"} Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.356923 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.357446 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b72f22392fc3adc615a4d5963b1705558ae485cd9f601d67db9bc14090b47b38" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.422346 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj"] Oct 09 10:34:13 crc kubenswrapper[4923]: E1009 10:34:13.423349 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0db9879d-8e0a-4d29-9168-caf8e16d5c3d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.423381 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="0db9879d-8e0a-4d29-9168-caf8e16d5c3d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.423572 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="0db9879d-8e0a-4d29-9168-caf8e16d5c3d" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.424368 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.428116 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.428114 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.428597 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.428810 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.438222 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj"] Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.576281 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.576653 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwb4f\" (UniqueName: \"kubernetes.io/projected/9f5ca5ba-bd14-40f8-aab3-094e1771a374-kube-api-access-rwb4f\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.576871 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.678985 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwb4f\" (UniqueName: \"kubernetes.io/projected/9f5ca5ba-bd14-40f8-aab3-094e1771a374-kube-api-access-rwb4f\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.679182 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.679316 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.683821 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.690405 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.698506 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwb4f\" (UniqueName: \"kubernetes.io/projected/9f5ca5ba-bd14-40f8-aab3-094e1771a374-kube-api-access-rwb4f\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:13 crc kubenswrapper[4923]: I1009 10:34:13.742833 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:14 crc kubenswrapper[4923]: I1009 10:34:14.095516 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj"] Oct 09 10:34:14 crc kubenswrapper[4923]: I1009 10:34:14.366673 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" event={"ID":"9f5ca5ba-bd14-40f8-aab3-094e1771a374","Type":"ContainerStarted","Data":"21ebddb9b7f4c3e84ee2ba1f4658ff6b32addd54f4f9e4ea4856958c46c63df4"} Oct 09 10:34:14 crc kubenswrapper[4923]: I1009 10:34:14.724240 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:34:15 crc kubenswrapper[4923]: I1009 10:34:15.378707 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" event={"ID":"9f5ca5ba-bd14-40f8-aab3-094e1771a374","Type":"ContainerStarted","Data":"52fa1ed0b40d689b747beaa7ce3213bf7ce43d82ed8ac44e9c75139f27c9891e"} Oct 09 10:34:15 crc kubenswrapper[4923]: I1009 10:34:15.407917 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" podStartSLOduration=1.793344965 podStartE2EDuration="2.407886033s" podCreationTimestamp="2025-10-09 10:34:13 +0000 UTC" firstStartedPulling="2025-10-09 10:34:14.103554014 +0000 UTC m=+1740.171735770" lastFinishedPulling="2025-10-09 10:34:14.718095082 +0000 UTC m=+1740.786276838" observedRunningTime="2025-10-09 10:34:15.399256243 +0000 UTC m=+1741.467438009" watchObservedRunningTime="2025-10-09 10:34:15.407886033 +0000 UTC m=+1741.476067799" Oct 09 10:34:15 crc kubenswrapper[4923]: I1009 10:34:15.601310 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:34:15 crc kubenswrapper[4923]: E1009 10:34:15.601641 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:34:16 crc kubenswrapper[4923]: I1009 10:34:16.033637 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-c481-account-create-mfwgm"] Oct 09 10:34:16 crc kubenswrapper[4923]: I1009 10:34:16.040546 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-c481-account-create-mfwgm"] Oct 09 10:34:16 crc kubenswrapper[4923]: I1009 10:34:16.613040 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6917f504-25e3-4991-a7cb-1f319586f9dd" path="/var/lib/kubelet/pods/6917f504-25e3-4991-a7cb-1f319586f9dd/volumes" Oct 09 10:34:18 crc kubenswrapper[4923]: I1009 10:34:18.029123 4923 scope.go:117] "RemoveContainer" containerID="1713cb05129fc81548f8522b7c9d09d32cc96f5a802add679b382b97fcb7743f" Oct 09 10:34:18 crc kubenswrapper[4923]: I1009 10:34:18.058528 4923 scope.go:117] "RemoveContainer" containerID="8917a51a7c0f6b85a9257086102be6807938116cc331605350a66bcfa7d2e5a0" Oct 09 10:34:18 crc kubenswrapper[4923]: I1009 10:34:18.141994 4923 scope.go:117] "RemoveContainer" containerID="4a342738a782fcf429cbbe261eb1c984df2b63b7a692620ca8d41c949a0c09fb" Oct 09 10:34:18 crc kubenswrapper[4923]: I1009 10:34:18.178829 4923 scope.go:117] "RemoveContainer" containerID="5cff3d46f1024cb5cf1326d24532df562280dd8442d8f50b896e203e5b64b080" Oct 09 10:34:18 crc kubenswrapper[4923]: I1009 10:34:18.247072 4923 scope.go:117] "RemoveContainer" containerID="23d8fa1b188a24014db4272f7656d4f568788e20dcdef3d0e243e575bb705d49" Oct 09 10:34:18 crc kubenswrapper[4923]: I1009 10:34:18.287051 4923 scope.go:117] "RemoveContainer" containerID="379ad10c6f26658939e5b9c00ad61fcc48f49cba5639a288227a541078da67cf" Oct 09 10:34:18 crc kubenswrapper[4923]: I1009 10:34:18.337117 4923 scope.go:117] "RemoveContainer" containerID="d3b3c2bbaec1c007eb581c20d7746c9adfd9bb779dcf06b99200d616a0a922f6" Oct 09 10:34:18 crc kubenswrapper[4923]: I1009 10:34:18.359321 4923 scope.go:117] "RemoveContainer" containerID="e3a247965a5127a1b0ba26eacf32079b5bde1889f0df43458415b5fbe19da36e" Oct 09 10:34:19 crc kubenswrapper[4923]: I1009 10:34:19.440675 4923 generic.go:334] "Generic (PLEG): container finished" podID="9f5ca5ba-bd14-40f8-aab3-094e1771a374" containerID="52fa1ed0b40d689b747beaa7ce3213bf7ce43d82ed8ac44e9c75139f27c9891e" exitCode=0 Oct 09 10:34:19 crc kubenswrapper[4923]: I1009 10:34:19.440793 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" event={"ID":"9f5ca5ba-bd14-40f8-aab3-094e1771a374","Type":"ContainerDied","Data":"52fa1ed0b40d689b747beaa7ce3213bf7ce43d82ed8ac44e9c75139f27c9891e"} Oct 09 10:34:20 crc kubenswrapper[4923]: I1009 10:34:20.932693 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.087417 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-inventory\") pod \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.087699 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-ssh-key\") pod \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.087901 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwb4f\" (UniqueName: \"kubernetes.io/projected/9f5ca5ba-bd14-40f8-aab3-094e1771a374-kube-api-access-rwb4f\") pod \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\" (UID: \"9f5ca5ba-bd14-40f8-aab3-094e1771a374\") " Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.097136 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f5ca5ba-bd14-40f8-aab3-094e1771a374-kube-api-access-rwb4f" (OuterVolumeSpecName: "kube-api-access-rwb4f") pod "9f5ca5ba-bd14-40f8-aab3-094e1771a374" (UID: "9f5ca5ba-bd14-40f8-aab3-094e1771a374"). InnerVolumeSpecName "kube-api-access-rwb4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.118224 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9f5ca5ba-bd14-40f8-aab3-094e1771a374" (UID: "9f5ca5ba-bd14-40f8-aab3-094e1771a374"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.122191 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-inventory" (OuterVolumeSpecName: "inventory") pod "9f5ca5ba-bd14-40f8-aab3-094e1771a374" (UID: "9f5ca5ba-bd14-40f8-aab3-094e1771a374"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.190615 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.190662 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9f5ca5ba-bd14-40f8-aab3-094e1771a374-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.190674 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwb4f\" (UniqueName: \"kubernetes.io/projected/9f5ca5ba-bd14-40f8-aab3-094e1771a374-kube-api-access-rwb4f\") on node \"crc\" DevicePath \"\"" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.480397 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" event={"ID":"9f5ca5ba-bd14-40f8-aab3-094e1771a374","Type":"ContainerDied","Data":"21ebddb9b7f4c3e84ee2ba1f4658ff6b32addd54f4f9e4ea4856958c46c63df4"} Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.480995 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21ebddb9b7f4c3e84ee2ba1f4658ff6b32addd54f4f9e4ea4856958c46c63df4" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.480794 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.546122 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b"] Oct 09 10:34:21 crc kubenswrapper[4923]: E1009 10:34:21.546622 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f5ca5ba-bd14-40f8-aab3-094e1771a374" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.546643 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f5ca5ba-bd14-40f8-aab3-094e1771a374" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.546855 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f5ca5ba-bd14-40f8-aab3-094e1771a374" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.547537 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.550814 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.551083 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.551364 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.551573 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.579361 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b"] Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.702240 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.702743 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fgvs\" (UniqueName: \"kubernetes.io/projected/d54c1adc-c247-4cd6-a97d-843ea434cad3-kube-api-access-6fgvs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.703169 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.805824 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fgvs\" (UniqueName: \"kubernetes.io/projected/d54c1adc-c247-4cd6-a97d-843ea434cad3-kube-api-access-6fgvs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.805897 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.806032 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.812192 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.812730 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.827609 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fgvs\" (UniqueName: \"kubernetes.io/projected/d54c1adc-c247-4cd6-a97d-843ea434cad3-kube-api-access-6fgvs\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:21 crc kubenswrapper[4923]: I1009 10:34:21.875968 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:34:22 crc kubenswrapper[4923]: I1009 10:34:22.474162 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b"] Oct 09 10:34:23 crc kubenswrapper[4923]: I1009 10:34:23.500088 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" event={"ID":"d54c1adc-c247-4cd6-a97d-843ea434cad3","Type":"ContainerStarted","Data":"40c4a05b7499f8c88b02b29ec6017cf46657afe8f377dbd3995a2317df9b2297"} Oct 09 10:34:23 crc kubenswrapper[4923]: I1009 10:34:23.500552 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" event={"ID":"d54c1adc-c247-4cd6-a97d-843ea434cad3","Type":"ContainerStarted","Data":"b75fb98f52cb6249792d02af19d0344bfd74715f3643d81c97a9876017e02915"} Oct 09 10:34:23 crc kubenswrapper[4923]: I1009 10:34:23.532792 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" podStartSLOduration=2.023514623 podStartE2EDuration="2.532741481s" podCreationTimestamp="2025-10-09 10:34:21 +0000 UTC" firstStartedPulling="2025-10-09 10:34:22.487307615 +0000 UTC m=+1748.555489371" lastFinishedPulling="2025-10-09 10:34:22.996534473 +0000 UTC m=+1749.064716229" observedRunningTime="2025-10-09 10:34:23.520543452 +0000 UTC m=+1749.588725208" watchObservedRunningTime="2025-10-09 10:34:23.532741481 +0000 UTC m=+1749.600923237" Oct 09 10:34:26 crc kubenswrapper[4923]: I1009 10:34:26.038114 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-ec7c-account-create-7ppmh"] Oct 09 10:34:26 crc kubenswrapper[4923]: I1009 10:34:26.051048 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-b76b-account-create-27zj7"] Oct 09 10:34:26 crc kubenswrapper[4923]: I1009 10:34:26.058876 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-ec7c-account-create-7ppmh"] Oct 09 10:34:26 crc kubenswrapper[4923]: I1009 10:34:26.066027 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-b76b-account-create-27zj7"] Oct 09 10:34:26 crc kubenswrapper[4923]: I1009 10:34:26.618010 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9637ebcc-b8d8-4f61-ae1b-372802dfa9ec" path="/var/lib/kubelet/pods/9637ebcc-b8d8-4f61-ae1b-372802dfa9ec/volumes" Oct 09 10:34:26 crc kubenswrapper[4923]: I1009 10:34:26.620190 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aeadce81-f642-4762-be0f-20650018847f" path="/var/lib/kubelet/pods/aeadce81-f642-4762-be0f-20650018847f/volumes" Oct 09 10:34:28 crc kubenswrapper[4923]: I1009 10:34:28.601587 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:34:28 crc kubenswrapper[4923]: E1009 10:34:28.602402 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:34:43 crc kubenswrapper[4923]: I1009 10:34:43.601563 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:34:43 crc kubenswrapper[4923]: E1009 10:34:43.602644 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:34:54 crc kubenswrapper[4923]: I1009 10:34:54.041084 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x5qkv"] Oct 09 10:34:54 crc kubenswrapper[4923]: I1009 10:34:54.050079 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-x5qkv"] Oct 09 10:34:54 crc kubenswrapper[4923]: I1009 10:34:54.612738 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31875d1d-5bd0-4a30-b82e-0113ad3cf237" path="/var/lib/kubelet/pods/31875d1d-5bd0-4a30-b82e-0113ad3cf237/volumes" Oct 09 10:34:55 crc kubenswrapper[4923]: I1009 10:34:55.601618 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:34:55 crc kubenswrapper[4923]: E1009 10:34:55.602170 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:35:09 crc kubenswrapper[4923]: I1009 10:35:09.602442 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:35:09 crc kubenswrapper[4923]: E1009 10:35:09.603811 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:35:18 crc kubenswrapper[4923]: I1009 10:35:18.538787 4923 scope.go:117] "RemoveContainer" containerID="51f2e52140e71d0bc17a819b33beaca75a330900540d0853813b9400663ba51e" Oct 09 10:35:18 crc kubenswrapper[4923]: I1009 10:35:18.605860 4923 scope.go:117] "RemoveContainer" containerID="2771c19824fd85f6e780d7e3e52832c155a101f65e3ba2b0d6e39419acfb2575" Oct 09 10:35:18 crc kubenswrapper[4923]: I1009 10:35:18.656408 4923 scope.go:117] "RemoveContainer" containerID="a0a8a6793f398684ad8ecdf254131a51037766c132b042b977951beae89080f1" Oct 09 10:35:19 crc kubenswrapper[4923]: I1009 10:35:19.036372 4923 generic.go:334] "Generic (PLEG): container finished" podID="d54c1adc-c247-4cd6-a97d-843ea434cad3" containerID="40c4a05b7499f8c88b02b29ec6017cf46657afe8f377dbd3995a2317df9b2297" exitCode=2 Oct 09 10:35:19 crc kubenswrapper[4923]: I1009 10:35:19.036492 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" event={"ID":"d54c1adc-c247-4cd6-a97d-843ea434cad3","Type":"ContainerDied","Data":"40c4a05b7499f8c88b02b29ec6017cf46657afe8f377dbd3995a2317df9b2297"} Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.052272 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-hkxgk"] Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.060371 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-hkxgk"] Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.444054 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.631440 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c4e9807-4e89-49d0-bea0-732464e07ca2" path="/var/lib/kubelet/pods/5c4e9807-4e89-49d0-bea0-732464e07ca2/volumes" Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.631572 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-ssh-key\") pod \"d54c1adc-c247-4cd6-a97d-843ea434cad3\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.631997 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-inventory\") pod \"d54c1adc-c247-4cd6-a97d-843ea434cad3\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.632088 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fgvs\" (UniqueName: \"kubernetes.io/projected/d54c1adc-c247-4cd6-a97d-843ea434cad3-kube-api-access-6fgvs\") pod \"d54c1adc-c247-4cd6-a97d-843ea434cad3\" (UID: \"d54c1adc-c247-4cd6-a97d-843ea434cad3\") " Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.637579 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d54c1adc-c247-4cd6-a97d-843ea434cad3-kube-api-access-6fgvs" (OuterVolumeSpecName: "kube-api-access-6fgvs") pod "d54c1adc-c247-4cd6-a97d-843ea434cad3" (UID: "d54c1adc-c247-4cd6-a97d-843ea434cad3"). InnerVolumeSpecName "kube-api-access-6fgvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.657887 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d54c1adc-c247-4cd6-a97d-843ea434cad3" (UID: "d54c1adc-c247-4cd6-a97d-843ea434cad3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.663083 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-inventory" (OuterVolumeSpecName: "inventory") pod "d54c1adc-c247-4cd6-a97d-843ea434cad3" (UID: "d54c1adc-c247-4cd6-a97d-843ea434cad3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.734911 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.734958 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d54c1adc-c247-4cd6-a97d-843ea434cad3-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:35:20 crc kubenswrapper[4923]: I1009 10:35:20.734970 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fgvs\" (UniqueName: \"kubernetes.io/projected/d54c1adc-c247-4cd6-a97d-843ea434cad3-kube-api-access-6fgvs\") on node \"crc\" DevicePath \"\"" Oct 09 10:35:21 crc kubenswrapper[4923]: I1009 10:35:21.063142 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" event={"ID":"d54c1adc-c247-4cd6-a97d-843ea434cad3","Type":"ContainerDied","Data":"b75fb98f52cb6249792d02af19d0344bfd74715f3643d81c97a9876017e02915"} Oct 09 10:35:21 crc kubenswrapper[4923]: I1009 10:35:21.063210 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b75fb98f52cb6249792d02af19d0344bfd74715f3643d81c97a9876017e02915" Oct 09 10:35:21 crc kubenswrapper[4923]: I1009 10:35:21.063265 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b" Oct 09 10:35:24 crc kubenswrapper[4923]: I1009 10:35:24.607001 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:35:24 crc kubenswrapper[4923]: E1009 10:35:24.607817 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:35:25 crc kubenswrapper[4923]: I1009 10:35:25.034965 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-78qxk"] Oct 09 10:35:25 crc kubenswrapper[4923]: I1009 10:35:25.046734 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-78qxk"] Oct 09 10:35:26 crc kubenswrapper[4923]: I1009 10:35:26.613313 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45dfedcc-9dbc-41e9-b3d8-e014b9409bf1" path="/var/lib/kubelet/pods/45dfedcc-9dbc-41e9-b3d8-e014b9409bf1/volumes" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.043368 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t"] Oct 09 10:35:28 crc kubenswrapper[4923]: E1009 10:35:28.044502 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d54c1adc-c247-4cd6-a97d-843ea434cad3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.044528 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d54c1adc-c247-4cd6-a97d-843ea434cad3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.044904 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d54c1adc-c247-4cd6-a97d-843ea434cad3" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.046184 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.052276 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.053188 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.054141 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.054192 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.060532 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t"] Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.081531 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bnxh\" (UniqueName: \"kubernetes.io/projected/9ba3fa71-9b42-4f4c-a1de-297d81d72269-kube-api-access-7bnxh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.081621 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.081654 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.183059 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bnxh\" (UniqueName: \"kubernetes.io/projected/9ba3fa71-9b42-4f4c-a1de-297d81d72269-kube-api-access-7bnxh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.183135 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.183169 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.190447 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.190930 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.213177 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bnxh\" (UniqueName: \"kubernetes.io/projected/9ba3fa71-9b42-4f4c-a1de-297d81d72269-kube-api-access-7bnxh\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.372887 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:35:28 crc kubenswrapper[4923]: I1009 10:35:28.958464 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t"] Oct 09 10:35:29 crc kubenswrapper[4923]: I1009 10:35:29.144175 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" event={"ID":"9ba3fa71-9b42-4f4c-a1de-297d81d72269","Type":"ContainerStarted","Data":"fd54ee36ce80e2a7a508f1c18b2a742d5ac9019ed584ee7c24635397c47a1a9e"} Oct 09 10:35:30 crc kubenswrapper[4923]: I1009 10:35:30.156299 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" event={"ID":"9ba3fa71-9b42-4f4c-a1de-297d81d72269","Type":"ContainerStarted","Data":"ec8e791db8671720b772fd554a378ba2dcce36ecdfb5d4e7343d97abbcb11535"} Oct 09 10:35:30 crc kubenswrapper[4923]: I1009 10:35:30.174450 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" podStartSLOduration=1.490739048 podStartE2EDuration="2.174427269s" podCreationTimestamp="2025-10-09 10:35:28 +0000 UTC" firstStartedPulling="2025-10-09 10:35:28.9642472 +0000 UTC m=+1815.032428956" lastFinishedPulling="2025-10-09 10:35:29.647935381 +0000 UTC m=+1815.716117177" observedRunningTime="2025-10-09 10:35:30.172467574 +0000 UTC m=+1816.240649340" watchObservedRunningTime="2025-10-09 10:35:30.174427269 +0000 UTC m=+1816.242609045" Oct 09 10:35:35 crc kubenswrapper[4923]: I1009 10:35:35.601274 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:35:36 crc kubenswrapper[4923]: I1009 10:35:36.211119 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"8b9ef1b7dba4296aa73497b9612a8d739ff91e430683907f0364c0ab200fbece"} Oct 09 10:36:04 crc kubenswrapper[4923]: I1009 10:36:04.056669 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-b2zqk"] Oct 09 10:36:04 crc kubenswrapper[4923]: I1009 10:36:04.064920 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-b2zqk"] Oct 09 10:36:04 crc kubenswrapper[4923]: I1009 10:36:04.614054 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf12e8f-fbe9-44dd-aff6-c8033ae56d45" path="/var/lib/kubelet/pods/daf12e8f-fbe9-44dd-aff6-c8033ae56d45/volumes" Oct 09 10:36:13 crc kubenswrapper[4923]: I1009 10:36:13.585088 4923 generic.go:334] "Generic (PLEG): container finished" podID="9ba3fa71-9b42-4f4c-a1de-297d81d72269" containerID="ec8e791db8671720b772fd554a378ba2dcce36ecdfb5d4e7343d97abbcb11535" exitCode=0 Oct 09 10:36:13 crc kubenswrapper[4923]: I1009 10:36:13.585158 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" event={"ID":"9ba3fa71-9b42-4f4c-a1de-297d81d72269","Type":"ContainerDied","Data":"ec8e791db8671720b772fd554a378ba2dcce36ecdfb5d4e7343d97abbcb11535"} Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.045632 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.067634 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-ssh-key\") pod \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.069628 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-inventory\") pod \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.069709 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bnxh\" (UniqueName: \"kubernetes.io/projected/9ba3fa71-9b42-4f4c-a1de-297d81d72269-kube-api-access-7bnxh\") pod \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\" (UID: \"9ba3fa71-9b42-4f4c-a1de-297d81d72269\") " Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.075266 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ba3fa71-9b42-4f4c-a1de-297d81d72269-kube-api-access-7bnxh" (OuterVolumeSpecName: "kube-api-access-7bnxh") pod "9ba3fa71-9b42-4f4c-a1de-297d81d72269" (UID: "9ba3fa71-9b42-4f4c-a1de-297d81d72269"). InnerVolumeSpecName "kube-api-access-7bnxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.101612 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-inventory" (OuterVolumeSpecName: "inventory") pod "9ba3fa71-9b42-4f4c-a1de-297d81d72269" (UID: "9ba3fa71-9b42-4f4c-a1de-297d81d72269"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.102555 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9ba3fa71-9b42-4f4c-a1de-297d81d72269" (UID: "9ba3fa71-9b42-4f4c-a1de-297d81d72269"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.179024 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bnxh\" (UniqueName: \"kubernetes.io/projected/9ba3fa71-9b42-4f4c-a1de-297d81d72269-kube-api-access-7bnxh\") on node \"crc\" DevicePath \"\"" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.179075 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.179092 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9ba3fa71-9b42-4f4c-a1de-297d81d72269-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.611077 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" event={"ID":"9ba3fa71-9b42-4f4c-a1de-297d81d72269","Type":"ContainerDied","Data":"fd54ee36ce80e2a7a508f1c18b2a742d5ac9019ed584ee7c24635397c47a1a9e"} Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.611137 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd54ee36ce80e2a7a508f1c18b2a742d5ac9019ed584ee7c24635397c47a1a9e" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.611534 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.698957 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w4cbw"] Oct 09 10:36:15 crc kubenswrapper[4923]: E1009 10:36:15.699669 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ba3fa71-9b42-4f4c-a1de-297d81d72269" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.699693 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ba3fa71-9b42-4f4c-a1de-297d81d72269" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.699907 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ba3fa71-9b42-4f4c-a1de-297d81d72269" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.700653 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.704114 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.706200 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.706317 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.706692 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.746450 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w4cbw"] Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.789897 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7cl9\" (UniqueName: \"kubernetes.io/projected/2d865110-029b-4108-8e59-cb3947a6269b-kube-api-access-n7cl9\") pod \"ssh-known-hosts-edpm-deployment-w4cbw\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.790076 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-w4cbw\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.790803 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-w4cbw\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.892688 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7cl9\" (UniqueName: \"kubernetes.io/projected/2d865110-029b-4108-8e59-cb3947a6269b-kube-api-access-n7cl9\") pod \"ssh-known-hosts-edpm-deployment-w4cbw\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.892834 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-w4cbw\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.892922 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-w4cbw\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.898163 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-w4cbw\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.908407 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-w4cbw\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:15 crc kubenswrapper[4923]: I1009 10:36:15.912126 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7cl9\" (UniqueName: \"kubernetes.io/projected/2d865110-029b-4108-8e59-cb3947a6269b-kube-api-access-n7cl9\") pod \"ssh-known-hosts-edpm-deployment-w4cbw\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:16 crc kubenswrapper[4923]: I1009 10:36:16.032773 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:16 crc kubenswrapper[4923]: I1009 10:36:16.580836 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w4cbw"] Oct 09 10:36:16 crc kubenswrapper[4923]: I1009 10:36:16.619820 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" event={"ID":"2d865110-029b-4108-8e59-cb3947a6269b","Type":"ContainerStarted","Data":"22ca2789c8b65df0886e84649b54db560e4a43bb8f717a72829dd3574b0a6cc6"} Oct 09 10:36:17 crc kubenswrapper[4923]: I1009 10:36:17.631946 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" event={"ID":"2d865110-029b-4108-8e59-cb3947a6269b","Type":"ContainerStarted","Data":"51ca72b4cdd99129b1f942fa19d543f88fb5745bedc7905827dbf9dfd25a2f6b"} Oct 09 10:36:17 crc kubenswrapper[4923]: I1009 10:36:17.651072 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" podStartSLOduration=2.166273635 podStartE2EDuration="2.651050882s" podCreationTimestamp="2025-10-09 10:36:15 +0000 UTC" firstStartedPulling="2025-10-09 10:36:16.587540284 +0000 UTC m=+1862.655722050" lastFinishedPulling="2025-10-09 10:36:17.072317551 +0000 UTC m=+1863.140499297" observedRunningTime="2025-10-09 10:36:17.647184474 +0000 UTC m=+1863.715366230" watchObservedRunningTime="2025-10-09 10:36:17.651050882 +0000 UTC m=+1863.719232638" Oct 09 10:36:18 crc kubenswrapper[4923]: I1009 10:36:18.768387 4923 scope.go:117] "RemoveContainer" containerID="9d7fbf539527d18ed86c2341708e436a44ee2c29f3a95f9f99732acbacd9422b" Oct 09 10:36:18 crc kubenswrapper[4923]: I1009 10:36:18.812239 4923 scope.go:117] "RemoveContainer" containerID="48f3f03b0b0e33fb4cb6af6c4d9de2cdc3336e631ffa235c713c4136c63a2637" Oct 09 10:36:18 crc kubenswrapper[4923]: I1009 10:36:18.893154 4923 scope.go:117] "RemoveContainer" containerID="05d08cf8c95399fd1938b6d73b8d94427366df1682385eba5b83ce0c1b34e101" Oct 09 10:36:24 crc kubenswrapper[4923]: I1009 10:36:24.701852 4923 generic.go:334] "Generic (PLEG): container finished" podID="2d865110-029b-4108-8e59-cb3947a6269b" containerID="51ca72b4cdd99129b1f942fa19d543f88fb5745bedc7905827dbf9dfd25a2f6b" exitCode=0 Oct 09 10:36:24 crc kubenswrapper[4923]: I1009 10:36:24.702007 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" event={"ID":"2d865110-029b-4108-8e59-cb3947a6269b","Type":"ContainerDied","Data":"51ca72b4cdd99129b1f942fa19d543f88fb5745bedc7905827dbf9dfd25a2f6b"} Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.169701 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.306795 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7cl9\" (UniqueName: \"kubernetes.io/projected/2d865110-029b-4108-8e59-cb3947a6269b-kube-api-access-n7cl9\") pod \"2d865110-029b-4108-8e59-cb3947a6269b\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.307026 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-inventory-0\") pod \"2d865110-029b-4108-8e59-cb3947a6269b\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.307195 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-ssh-key-openstack-edpm-ipam\") pod \"2d865110-029b-4108-8e59-cb3947a6269b\" (UID: \"2d865110-029b-4108-8e59-cb3947a6269b\") " Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.314378 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d865110-029b-4108-8e59-cb3947a6269b-kube-api-access-n7cl9" (OuterVolumeSpecName: "kube-api-access-n7cl9") pod "2d865110-029b-4108-8e59-cb3947a6269b" (UID: "2d865110-029b-4108-8e59-cb3947a6269b"). InnerVolumeSpecName "kube-api-access-n7cl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.339309 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "2d865110-029b-4108-8e59-cb3947a6269b" (UID: "2d865110-029b-4108-8e59-cb3947a6269b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.339906 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "2d865110-029b-4108-8e59-cb3947a6269b" (UID: "2d865110-029b-4108-8e59-cb3947a6269b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.410225 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.410291 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7cl9\" (UniqueName: \"kubernetes.io/projected/2d865110-029b-4108-8e59-cb3947a6269b-kube-api-access-n7cl9\") on node \"crc\" DevicePath \"\"" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.410304 4923 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/2d865110-029b-4108-8e59-cb3947a6269b-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.732378 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" event={"ID":"2d865110-029b-4108-8e59-cb3947a6269b","Type":"ContainerDied","Data":"22ca2789c8b65df0886e84649b54db560e4a43bb8f717a72829dd3574b0a6cc6"} Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.732433 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w4cbw" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.732449 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22ca2789c8b65df0886e84649b54db560e4a43bb8f717a72829dd3574b0a6cc6" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.802864 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h"] Oct 09 10:36:26 crc kubenswrapper[4923]: E1009 10:36:26.803400 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d865110-029b-4108-8e59-cb3947a6269b" containerName="ssh-known-hosts-edpm-deployment" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.803430 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d865110-029b-4108-8e59-cb3947a6269b" containerName="ssh-known-hosts-edpm-deployment" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.803673 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d865110-029b-4108-8e59-cb3947a6269b" containerName="ssh-known-hosts-edpm-deployment" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.804555 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.811034 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.812218 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.812485 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.813433 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h"] Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.815361 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.920426 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxdfc\" (UniqueName: \"kubernetes.io/projected/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-kube-api-access-vxdfc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2279h\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.920517 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2279h\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:26 crc kubenswrapper[4923]: I1009 10:36:26.920751 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2279h\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:27 crc kubenswrapper[4923]: I1009 10:36:27.024576 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2279h\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:27 crc kubenswrapper[4923]: I1009 10:36:27.025579 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2279h\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:27 crc kubenswrapper[4923]: I1009 10:36:27.025818 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxdfc\" (UniqueName: \"kubernetes.io/projected/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-kube-api-access-vxdfc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2279h\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:27 crc kubenswrapper[4923]: I1009 10:36:27.030637 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2279h\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:27 crc kubenswrapper[4923]: I1009 10:36:27.042360 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2279h\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:27 crc kubenswrapper[4923]: I1009 10:36:27.047576 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxdfc\" (UniqueName: \"kubernetes.io/projected/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-kube-api-access-vxdfc\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-2279h\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:27 crc kubenswrapper[4923]: I1009 10:36:27.121370 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:27 crc kubenswrapper[4923]: I1009 10:36:27.680302 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h"] Oct 09 10:36:27 crc kubenswrapper[4923]: I1009 10:36:27.743901 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" event={"ID":"cc3626c2-ad33-40b5-a56d-1e4ea07696f2","Type":"ContainerStarted","Data":"92d7817cba4cd718e3b21d289405e74c03f3e918c82bc66d97ec8b1b5d2748bf"} Oct 09 10:36:28 crc kubenswrapper[4923]: I1009 10:36:28.757164 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" event={"ID":"cc3626c2-ad33-40b5-a56d-1e4ea07696f2","Type":"ContainerStarted","Data":"29aac3997105a778d531276d0b53f750c56e70f4235ed3bafa5a15890e43455a"} Oct 09 10:36:37 crc kubenswrapper[4923]: I1009 10:36:37.843196 4923 generic.go:334] "Generic (PLEG): container finished" podID="cc3626c2-ad33-40b5-a56d-1e4ea07696f2" containerID="29aac3997105a778d531276d0b53f750c56e70f4235ed3bafa5a15890e43455a" exitCode=0 Oct 09 10:36:37 crc kubenswrapper[4923]: I1009 10:36:37.843265 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" event={"ID":"cc3626c2-ad33-40b5-a56d-1e4ea07696f2","Type":"ContainerDied","Data":"29aac3997105a778d531276d0b53f750c56e70f4235ed3bafa5a15890e43455a"} Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.313114 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.406956 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-inventory\") pod \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.407108 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-ssh-key\") pod \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.407323 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxdfc\" (UniqueName: \"kubernetes.io/projected/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-kube-api-access-vxdfc\") pod \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\" (UID: \"cc3626c2-ad33-40b5-a56d-1e4ea07696f2\") " Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.418225 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-kube-api-access-vxdfc" (OuterVolumeSpecName: "kube-api-access-vxdfc") pod "cc3626c2-ad33-40b5-a56d-1e4ea07696f2" (UID: "cc3626c2-ad33-40b5-a56d-1e4ea07696f2"). InnerVolumeSpecName "kube-api-access-vxdfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.441756 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-inventory" (OuterVolumeSpecName: "inventory") pod "cc3626c2-ad33-40b5-a56d-1e4ea07696f2" (UID: "cc3626c2-ad33-40b5-a56d-1e4ea07696f2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.441894 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cc3626c2-ad33-40b5-a56d-1e4ea07696f2" (UID: "cc3626c2-ad33-40b5-a56d-1e4ea07696f2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.510306 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.510352 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.510366 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxdfc\" (UniqueName: \"kubernetes.io/projected/cc3626c2-ad33-40b5-a56d-1e4ea07696f2-kube-api-access-vxdfc\") on node \"crc\" DevicePath \"\"" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.867017 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.869811 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h" event={"ID":"cc3626c2-ad33-40b5-a56d-1e4ea07696f2","Type":"ContainerDied","Data":"92d7817cba4cd718e3b21d289405e74c03f3e918c82bc66d97ec8b1b5d2748bf"} Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.869889 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92d7817cba4cd718e3b21d289405e74c03f3e918c82bc66d97ec8b1b5d2748bf" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.965528 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6"] Oct 09 10:36:39 crc kubenswrapper[4923]: E1009 10:36:39.966123 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc3626c2-ad33-40b5-a56d-1e4ea07696f2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.966149 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc3626c2-ad33-40b5-a56d-1e4ea07696f2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.966431 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc3626c2-ad33-40b5-a56d-1e4ea07696f2" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.967314 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.978013 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.978101 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.978111 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.978375 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:36:39 crc kubenswrapper[4923]: I1009 10:36:39.982721 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6"] Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.123893 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9n5x\" (UniqueName: \"kubernetes.io/projected/1d78a009-a62b-41e8-8384-3c1f668ced70-kube-api-access-f9n5x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.123981 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.124457 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.226703 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.226872 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9n5x\" (UniqueName: \"kubernetes.io/projected/1d78a009-a62b-41e8-8384-3c1f668ced70-kube-api-access-f9n5x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.226959 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.237111 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.237178 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.244805 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9n5x\" (UniqueName: \"kubernetes.io/projected/1d78a009-a62b-41e8-8384-3c1f668ced70-kube-api-access-f9n5x\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.292961 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:36:40 crc kubenswrapper[4923]: I1009 10:36:40.935680 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6"] Oct 09 10:36:41 crc kubenswrapper[4923]: I1009 10:36:41.902957 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" event={"ID":"1d78a009-a62b-41e8-8384-3c1f668ced70","Type":"ContainerStarted","Data":"eb6d1b4698184a8ae85ee523702ffcff8b4686aca4961dddc1f69401ab693b7d"} Oct 09 10:36:42 crc kubenswrapper[4923]: I1009 10:36:42.917680 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" event={"ID":"1d78a009-a62b-41e8-8384-3c1f668ced70","Type":"ContainerStarted","Data":"7513b828ccc587c4bd7f26721212b18413bd6398bbaae3f9f439ab9c1b356349"} Oct 09 10:36:42 crc kubenswrapper[4923]: I1009 10:36:42.943927 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" podStartSLOduration=3.166986272 podStartE2EDuration="3.943902463s" podCreationTimestamp="2025-10-09 10:36:39 +0000 UTC" firstStartedPulling="2025-10-09 10:36:40.941065494 +0000 UTC m=+1887.009247250" lastFinishedPulling="2025-10-09 10:36:41.717981685 +0000 UTC m=+1887.786163441" observedRunningTime="2025-10-09 10:36:42.93370964 +0000 UTC m=+1889.001891406" watchObservedRunningTime="2025-10-09 10:36:42.943902463 +0000 UTC m=+1889.012084219" Oct 09 10:37:54 crc kubenswrapper[4923]: I1009 10:37:54.599405 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:37:54 crc kubenswrapper[4923]: I1009 10:37:54.600052 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:37:57 crc kubenswrapper[4923]: I1009 10:37:57.696974 4923 generic.go:334] "Generic (PLEG): container finished" podID="1d78a009-a62b-41e8-8384-3c1f668ced70" containerID="7513b828ccc587c4bd7f26721212b18413bd6398bbaae3f9f439ab9c1b356349" exitCode=0 Oct 09 10:37:57 crc kubenswrapper[4923]: I1009 10:37:57.697064 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" event={"ID":"1d78a009-a62b-41e8-8384-3c1f668ced70","Type":"ContainerDied","Data":"7513b828ccc587c4bd7f26721212b18413bd6398bbaae3f9f439ab9c1b356349"} Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.180634 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.289595 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-inventory\") pod \"1d78a009-a62b-41e8-8384-3c1f668ced70\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.289884 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-ssh-key\") pod \"1d78a009-a62b-41e8-8384-3c1f668ced70\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.290093 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9n5x\" (UniqueName: \"kubernetes.io/projected/1d78a009-a62b-41e8-8384-3c1f668ced70-kube-api-access-f9n5x\") pod \"1d78a009-a62b-41e8-8384-3c1f668ced70\" (UID: \"1d78a009-a62b-41e8-8384-3c1f668ced70\") " Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.298144 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d78a009-a62b-41e8-8384-3c1f668ced70-kube-api-access-f9n5x" (OuterVolumeSpecName: "kube-api-access-f9n5x") pod "1d78a009-a62b-41e8-8384-3c1f668ced70" (UID: "1d78a009-a62b-41e8-8384-3c1f668ced70"). InnerVolumeSpecName "kube-api-access-f9n5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.327055 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-inventory" (OuterVolumeSpecName: "inventory") pod "1d78a009-a62b-41e8-8384-3c1f668ced70" (UID: "1d78a009-a62b-41e8-8384-3c1f668ced70"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.327254 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1d78a009-a62b-41e8-8384-3c1f668ced70" (UID: "1d78a009-a62b-41e8-8384-3c1f668ced70"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.394174 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9n5x\" (UniqueName: \"kubernetes.io/projected/1d78a009-a62b-41e8-8384-3c1f668ced70-kube-api-access-f9n5x\") on node \"crc\" DevicePath \"\"" Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.394218 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.394229 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1d78a009-a62b-41e8-8384-3c1f668ced70-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.724323 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" event={"ID":"1d78a009-a62b-41e8-8384-3c1f668ced70","Type":"ContainerDied","Data":"eb6d1b4698184a8ae85ee523702ffcff8b4686aca4961dddc1f69401ab693b7d"} Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.724383 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb6d1b4698184a8ae85ee523702ffcff8b4686aca4961dddc1f69401ab693b7d" Oct 09 10:37:59 crc kubenswrapper[4923]: I1009 10:37:59.724492 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6" Oct 09 10:38:24 crc kubenswrapper[4923]: I1009 10:38:24.599889 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:38:24 crc kubenswrapper[4923]: I1009 10:38:24.600937 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:38:54 crc kubenswrapper[4923]: I1009 10:38:54.599656 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:38:54 crc kubenswrapper[4923]: I1009 10:38:54.600497 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:38:54 crc kubenswrapper[4923]: I1009 10:38:54.600646 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:38:54 crc kubenswrapper[4923]: I1009 10:38:54.607621 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8b9ef1b7dba4296aa73497b9612a8d739ff91e430683907f0364c0ab200fbece"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:38:54 crc kubenswrapper[4923]: I1009 10:38:54.607706 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://8b9ef1b7dba4296aa73497b9612a8d739ff91e430683907f0364c0ab200fbece" gracePeriod=600 Oct 09 10:38:55 crc kubenswrapper[4923]: I1009 10:38:55.240363 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="8b9ef1b7dba4296aa73497b9612a8d739ff91e430683907f0364c0ab200fbece" exitCode=0 Oct 09 10:38:55 crc kubenswrapper[4923]: I1009 10:38:55.240441 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"8b9ef1b7dba4296aa73497b9612a8d739ff91e430683907f0364c0ab200fbece"} Oct 09 10:38:55 crc kubenswrapper[4923]: I1009 10:38:55.241215 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391"} Oct 09 10:38:55 crc kubenswrapper[4923]: I1009 10:38:55.241245 4923 scope.go:117] "RemoveContainer" containerID="a053b9f5b068c4c1fbb91c1a3f147f1bb631009a4d00b527d6847afe08306487" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.407769 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-84ks5"] Oct 09 10:40:06 crc kubenswrapper[4923]: E1009 10:40:06.408923 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d78a009-a62b-41e8-8384-3c1f668ced70" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.408946 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d78a009-a62b-41e8-8384-3c1f668ced70" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.409132 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d78a009-a62b-41e8-8384-3c1f668ced70" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.410947 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.424034 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84ks5"] Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.536485 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-utilities\") pod \"community-operators-84ks5\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.537147 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-catalog-content\") pod \"community-operators-84ks5\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.537211 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqw6j\" (UniqueName: \"kubernetes.io/projected/d70d4cf2-1ab1-4025-864d-37dd1eae658c-kube-api-access-jqw6j\") pod \"community-operators-84ks5\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.638957 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-catalog-content\") pod \"community-operators-84ks5\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.639030 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqw6j\" (UniqueName: \"kubernetes.io/projected/d70d4cf2-1ab1-4025-864d-37dd1eae658c-kube-api-access-jqw6j\") pod \"community-operators-84ks5\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.639173 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-utilities\") pod \"community-operators-84ks5\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.639499 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-catalog-content\") pod \"community-operators-84ks5\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.639804 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-utilities\") pod \"community-operators-84ks5\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.668408 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqw6j\" (UniqueName: \"kubernetes.io/projected/d70d4cf2-1ab1-4025-864d-37dd1eae658c-kube-api-access-jqw6j\") pod \"community-operators-84ks5\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:06 crc kubenswrapper[4923]: I1009 10:40:06.741888 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:07 crc kubenswrapper[4923]: I1009 10:40:07.290374 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-84ks5"] Oct 09 10:40:07 crc kubenswrapper[4923]: I1009 10:40:07.981411 4923 generic.go:334] "Generic (PLEG): container finished" podID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerID="3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d" exitCode=0 Oct 09 10:40:07 crc kubenswrapper[4923]: I1009 10:40:07.981461 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84ks5" event={"ID":"d70d4cf2-1ab1-4025-864d-37dd1eae658c","Type":"ContainerDied","Data":"3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d"} Oct 09 10:40:07 crc kubenswrapper[4923]: I1009 10:40:07.981844 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84ks5" event={"ID":"d70d4cf2-1ab1-4025-864d-37dd1eae658c","Type":"ContainerStarted","Data":"f154a742d48f4a19a370bf4366550ab8873e8857b5355cc0a10a21f6766a9564"} Oct 09 10:40:07 crc kubenswrapper[4923]: I1009 10:40:07.985580 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 10:40:10 crc kubenswrapper[4923]: I1009 10:40:10.020892 4923 generic.go:334] "Generic (PLEG): container finished" podID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerID="f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737" exitCode=0 Oct 09 10:40:10 crc kubenswrapper[4923]: I1009 10:40:10.021039 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84ks5" event={"ID":"d70d4cf2-1ab1-4025-864d-37dd1eae658c","Type":"ContainerDied","Data":"f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737"} Oct 09 10:40:11 crc kubenswrapper[4923]: I1009 10:40:11.791376 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f8xkh"] Oct 09 10:40:11 crc kubenswrapper[4923]: I1009 10:40:11.794858 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:11 crc kubenswrapper[4923]: I1009 10:40:11.808715 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8xkh"] Oct 09 10:40:11 crc kubenswrapper[4923]: I1009 10:40:11.947375 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2fxf\" (UniqueName: \"kubernetes.io/projected/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-kube-api-access-b2fxf\") pod \"redhat-marketplace-f8xkh\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:11 crc kubenswrapper[4923]: I1009 10:40:11.947436 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-utilities\") pod \"redhat-marketplace-f8xkh\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:11 crc kubenswrapper[4923]: I1009 10:40:11.947541 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-catalog-content\") pod \"redhat-marketplace-f8xkh\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.042683 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84ks5" event={"ID":"d70d4cf2-1ab1-4025-864d-37dd1eae658c","Type":"ContainerStarted","Data":"f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62"} Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.049176 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2fxf\" (UniqueName: \"kubernetes.io/projected/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-kube-api-access-b2fxf\") pod \"redhat-marketplace-f8xkh\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.049250 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-utilities\") pod \"redhat-marketplace-f8xkh\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.049364 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-catalog-content\") pod \"redhat-marketplace-f8xkh\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.049903 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-utilities\") pod \"redhat-marketplace-f8xkh\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.050005 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-catalog-content\") pod \"redhat-marketplace-f8xkh\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.072717 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2fxf\" (UniqueName: \"kubernetes.io/projected/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-kube-api-access-b2fxf\") pod \"redhat-marketplace-f8xkh\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.075806 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-84ks5" podStartSLOduration=3.233265479 podStartE2EDuration="6.075731451s" podCreationTimestamp="2025-10-09 10:40:06 +0000 UTC" firstStartedPulling="2025-10-09 10:40:07.985325049 +0000 UTC m=+2094.053506805" lastFinishedPulling="2025-10-09 10:40:10.827791001 +0000 UTC m=+2096.895972777" observedRunningTime="2025-10-09 10:40:12.06277346 +0000 UTC m=+2098.130955226" watchObservedRunningTime="2025-10-09 10:40:12.075731451 +0000 UTC m=+2098.143913207" Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.117422 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:12 crc kubenswrapper[4923]: I1009 10:40:12.469820 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8xkh"] Oct 09 10:40:12 crc kubenswrapper[4923]: W1009 10:40:12.478535 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c3e1e21_fe32_403f_b4be_aaadc9379e9d.slice/crio-7bb1026daf496f1e082355f2e835c8f4579a16f9e11b338e97dc65baa83f69c1 WatchSource:0}: Error finding container 7bb1026daf496f1e082355f2e835c8f4579a16f9e11b338e97dc65baa83f69c1: Status 404 returned error can't find the container with id 7bb1026daf496f1e082355f2e835c8f4579a16f9e11b338e97dc65baa83f69c1 Oct 09 10:40:13 crc kubenswrapper[4923]: I1009 10:40:13.059248 4923 generic.go:334] "Generic (PLEG): container finished" podID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerID="53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1" exitCode=0 Oct 09 10:40:13 crc kubenswrapper[4923]: I1009 10:40:13.059424 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8xkh" event={"ID":"3c3e1e21-fe32-403f-b4be-aaadc9379e9d","Type":"ContainerDied","Data":"53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1"} Oct 09 10:40:13 crc kubenswrapper[4923]: I1009 10:40:13.059872 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8xkh" event={"ID":"3c3e1e21-fe32-403f-b4be-aaadc9379e9d","Type":"ContainerStarted","Data":"7bb1026daf496f1e082355f2e835c8f4579a16f9e11b338e97dc65baa83f69c1"} Oct 09 10:40:14 crc kubenswrapper[4923]: I1009 10:40:14.072250 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8xkh" event={"ID":"3c3e1e21-fe32-403f-b4be-aaadc9379e9d","Type":"ContainerStarted","Data":"9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e"} Oct 09 10:40:15 crc kubenswrapper[4923]: I1009 10:40:15.083323 4923 generic.go:334] "Generic (PLEG): container finished" podID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerID="9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e" exitCode=0 Oct 09 10:40:15 crc kubenswrapper[4923]: I1009 10:40:15.083378 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8xkh" event={"ID":"3c3e1e21-fe32-403f-b4be-aaadc9379e9d","Type":"ContainerDied","Data":"9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e"} Oct 09 10:40:16 crc kubenswrapper[4923]: I1009 10:40:16.098350 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8xkh" event={"ID":"3c3e1e21-fe32-403f-b4be-aaadc9379e9d","Type":"ContainerStarted","Data":"d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba"} Oct 09 10:40:16 crc kubenswrapper[4923]: I1009 10:40:16.125189 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f8xkh" podStartSLOduration=2.662278621 podStartE2EDuration="5.125163975s" podCreationTimestamp="2025-10-09 10:40:11 +0000 UTC" firstStartedPulling="2025-10-09 10:40:13.061745532 +0000 UTC m=+2099.129927288" lastFinishedPulling="2025-10-09 10:40:15.524630876 +0000 UTC m=+2101.592812642" observedRunningTime="2025-10-09 10:40:16.120635909 +0000 UTC m=+2102.188817685" watchObservedRunningTime="2025-10-09 10:40:16.125163975 +0000 UTC m=+2102.193345731" Oct 09 10:40:16 crc kubenswrapper[4923]: I1009 10:40:16.742532 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:16 crc kubenswrapper[4923]: I1009 10:40:16.742995 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:16 crc kubenswrapper[4923]: I1009 10:40:16.794126 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:17 crc kubenswrapper[4923]: I1009 10:40:17.192692 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:18 crc kubenswrapper[4923]: I1009 10:40:18.983853 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-84ks5"] Oct 09 10:40:19 crc kubenswrapper[4923]: I1009 10:40:19.135488 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-84ks5" podUID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerName="registry-server" containerID="cri-o://f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62" gracePeriod=2 Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.090856 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.145189 4923 generic.go:334] "Generic (PLEG): container finished" podID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerID="f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62" exitCode=0 Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.145282 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84ks5" event={"ID":"d70d4cf2-1ab1-4025-864d-37dd1eae658c","Type":"ContainerDied","Data":"f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62"} Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.145322 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-84ks5" event={"ID":"d70d4cf2-1ab1-4025-864d-37dd1eae658c","Type":"ContainerDied","Data":"f154a742d48f4a19a370bf4366550ab8873e8857b5355cc0a10a21f6766a9564"} Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.145345 4923 scope.go:117] "RemoveContainer" containerID="f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.145499 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-84ks5" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.165200 4923 scope.go:117] "RemoveContainer" containerID="f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.183718 4923 scope.go:117] "RemoveContainer" containerID="3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.231746 4923 scope.go:117] "RemoveContainer" containerID="f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62" Oct 09 10:40:20 crc kubenswrapper[4923]: E1009 10:40:20.232213 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62\": container with ID starting with f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62 not found: ID does not exist" containerID="f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.232259 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62"} err="failed to get container status \"f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62\": rpc error: code = NotFound desc = could not find container \"f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62\": container with ID starting with f046146409027bd77fde7dff88158b214136909789ff5cf2a6014511ca3a9f62 not found: ID does not exist" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.232287 4923 scope.go:117] "RemoveContainer" containerID="f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737" Oct 09 10:40:20 crc kubenswrapper[4923]: E1009 10:40:20.232629 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737\": container with ID starting with f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737 not found: ID does not exist" containerID="f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.232719 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737"} err="failed to get container status \"f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737\": rpc error: code = NotFound desc = could not find container \"f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737\": container with ID starting with f9445acddb61292446eb92afcb75b4d20fc618f150aa25875bc6c86c89271737 not found: ID does not exist" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.232801 4923 scope.go:117] "RemoveContainer" containerID="3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d" Oct 09 10:40:20 crc kubenswrapper[4923]: E1009 10:40:20.233171 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d\": container with ID starting with 3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d not found: ID does not exist" containerID="3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.233199 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d"} err="failed to get container status \"3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d\": rpc error: code = NotFound desc = could not find container \"3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d\": container with ID starting with 3b0eef408bd9afd1fd55be2446872fd4cecd7490fcdfa63bece8d1a653d5c51d not found: ID does not exist" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.238037 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqw6j\" (UniqueName: \"kubernetes.io/projected/d70d4cf2-1ab1-4025-864d-37dd1eae658c-kube-api-access-jqw6j\") pod \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.238206 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-utilities\") pod \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.238368 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-catalog-content\") pod \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\" (UID: \"d70d4cf2-1ab1-4025-864d-37dd1eae658c\") " Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.239573 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-utilities" (OuterVolumeSpecName: "utilities") pod "d70d4cf2-1ab1-4025-864d-37dd1eae658c" (UID: "d70d4cf2-1ab1-4025-864d-37dd1eae658c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.243879 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d70d4cf2-1ab1-4025-864d-37dd1eae658c-kube-api-access-jqw6j" (OuterVolumeSpecName: "kube-api-access-jqw6j") pod "d70d4cf2-1ab1-4025-864d-37dd1eae658c" (UID: "d70d4cf2-1ab1-4025-864d-37dd1eae658c"). InnerVolumeSpecName "kube-api-access-jqw6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.291010 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d70d4cf2-1ab1-4025-864d-37dd1eae658c" (UID: "d70d4cf2-1ab1-4025-864d-37dd1eae658c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.341379 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.342018 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d70d4cf2-1ab1-4025-864d-37dd1eae658c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.342099 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqw6j\" (UniqueName: \"kubernetes.io/projected/d70d4cf2-1ab1-4025-864d-37dd1eae658c-kube-api-access-jqw6j\") on node \"crc\" DevicePath \"\"" Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.489677 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-84ks5"] Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.495577 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-84ks5"] Oct 09 10:40:20 crc kubenswrapper[4923]: I1009 10:40:20.614382 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" path="/var/lib/kubelet/pods/d70d4cf2-1ab1-4025-864d-37dd1eae658c/volumes" Oct 09 10:40:22 crc kubenswrapper[4923]: I1009 10:40:22.117681 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:22 crc kubenswrapper[4923]: I1009 10:40:22.118094 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:22 crc kubenswrapper[4923]: I1009 10:40:22.174252 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:22 crc kubenswrapper[4923]: I1009 10:40:22.226264 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:23 crc kubenswrapper[4923]: I1009 10:40:23.386583 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8xkh"] Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.189638 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f8xkh" podUID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerName="registry-server" containerID="cri-o://d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba" gracePeriod=2 Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.695669 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.837155 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-utilities\") pod \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.837351 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2fxf\" (UniqueName: \"kubernetes.io/projected/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-kube-api-access-b2fxf\") pod \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.837476 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-catalog-content\") pod \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\" (UID: \"3c3e1e21-fe32-403f-b4be-aaadc9379e9d\") " Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.838617 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-utilities" (OuterVolumeSpecName: "utilities") pod "3c3e1e21-fe32-403f-b4be-aaadc9379e9d" (UID: "3c3e1e21-fe32-403f-b4be-aaadc9379e9d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.846265 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-kube-api-access-b2fxf" (OuterVolumeSpecName: "kube-api-access-b2fxf") pod "3c3e1e21-fe32-403f-b4be-aaadc9379e9d" (UID: "3c3e1e21-fe32-403f-b4be-aaadc9379e9d"). InnerVolumeSpecName "kube-api-access-b2fxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.854055 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3c3e1e21-fe32-403f-b4be-aaadc9379e9d" (UID: "3c3e1e21-fe32-403f-b4be-aaadc9379e9d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.942496 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2fxf\" (UniqueName: \"kubernetes.io/projected/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-kube-api-access-b2fxf\") on node \"crc\" DevicePath \"\"" Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.943053 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:40:24 crc kubenswrapper[4923]: I1009 10:40:24.943067 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3c3e1e21-fe32-403f-b4be-aaadc9379e9d-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.203470 4923 generic.go:334] "Generic (PLEG): container finished" podID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerID="d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba" exitCode=0 Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.203525 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8xkh" event={"ID":"3c3e1e21-fe32-403f-b4be-aaadc9379e9d","Type":"ContainerDied","Data":"d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba"} Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.203561 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f8xkh" event={"ID":"3c3e1e21-fe32-403f-b4be-aaadc9379e9d","Type":"ContainerDied","Data":"7bb1026daf496f1e082355f2e835c8f4579a16f9e11b338e97dc65baa83f69c1"} Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.203581 4923 scope.go:117] "RemoveContainer" containerID="d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.204116 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f8xkh" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.228315 4923 scope.go:117] "RemoveContainer" containerID="9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.244697 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8xkh"] Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.253365 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f8xkh"] Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.262248 4923 scope.go:117] "RemoveContainer" containerID="53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.318356 4923 scope.go:117] "RemoveContainer" containerID="d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba" Oct 09 10:40:25 crc kubenswrapper[4923]: E1009 10:40:25.319139 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba\": container with ID starting with d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba not found: ID does not exist" containerID="d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.319176 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba"} err="failed to get container status \"d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba\": rpc error: code = NotFound desc = could not find container \"d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba\": container with ID starting with d8bc4befdf823674faa4466ae7c616adf35eab88c243a675d2d2280ef3b031ba not found: ID does not exist" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.319201 4923 scope.go:117] "RemoveContainer" containerID="9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e" Oct 09 10:40:25 crc kubenswrapper[4923]: E1009 10:40:25.319631 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e\": container with ID starting with 9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e not found: ID does not exist" containerID="9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.319691 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e"} err="failed to get container status \"9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e\": rpc error: code = NotFound desc = could not find container \"9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e\": container with ID starting with 9c8c135a402431ef9d9377a232b89e1932043d1f1e7f9f5e201e19d9c44c0a5e not found: ID does not exist" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.319732 4923 scope.go:117] "RemoveContainer" containerID="53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1" Oct 09 10:40:25 crc kubenswrapper[4923]: E1009 10:40:25.320372 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1\": container with ID starting with 53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1 not found: ID does not exist" containerID="53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1" Oct 09 10:40:25 crc kubenswrapper[4923]: I1009 10:40:25.320417 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1"} err="failed to get container status \"53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1\": rpc error: code = NotFound desc = could not find container \"53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1\": container with ID starting with 53997d98a3486f6b8a82ce77da44bb4ea1587715e296e375fa99be2ca474b2f1 not found: ID does not exist" Oct 09 10:40:26 crc kubenswrapper[4923]: I1009 10:40:26.615219 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" path="/var/lib/kubelet/pods/3c3e1e21-fe32-403f-b4be-aaadc9379e9d/volumes" Oct 09 10:40:54 crc kubenswrapper[4923]: I1009 10:40:54.599938 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:40:54 crc kubenswrapper[4923]: I1009 10:40:54.601396 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:41:24 crc kubenswrapper[4923]: I1009 10:41:24.599704 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:41:24 crc kubenswrapper[4923]: I1009 10:41:24.600636 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:41:54 crc kubenswrapper[4923]: I1009 10:41:54.600347 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:41:54 crc kubenswrapper[4923]: I1009 10:41:54.601160 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:41:54 crc kubenswrapper[4923]: I1009 10:41:54.612738 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:41:54 crc kubenswrapper[4923]: I1009 10:41:54.613777 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:41:54 crc kubenswrapper[4923]: I1009 10:41:54.613855 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" gracePeriod=600 Oct 09 10:41:54 crc kubenswrapper[4923]: E1009 10:41:54.747890 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:41:55 crc kubenswrapper[4923]: I1009 10:41:55.064324 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" exitCode=0 Oct 09 10:41:55 crc kubenswrapper[4923]: I1009 10:41:55.064380 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391"} Oct 09 10:41:55 crc kubenswrapper[4923]: I1009 10:41:55.064436 4923 scope.go:117] "RemoveContainer" containerID="8b9ef1b7dba4296aa73497b9612a8d739ff91e430683907f0364c0ab200fbece" Oct 09 10:41:55 crc kubenswrapper[4923]: I1009 10:41:55.065368 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:41:55 crc kubenswrapper[4923]: E1009 10:41:55.065739 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:42:09 crc kubenswrapper[4923]: I1009 10:42:09.603087 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:42:09 crc kubenswrapper[4923]: E1009 10:42:09.604152 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:42:21 crc kubenswrapper[4923]: I1009 10:42:21.602551 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:42:21 crc kubenswrapper[4923]: E1009 10:42:21.604450 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:42:34 crc kubenswrapper[4923]: I1009 10:42:34.610250 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:42:34 crc kubenswrapper[4923]: E1009 10:42:34.611352 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:42:45 crc kubenswrapper[4923]: I1009 10:42:45.601736 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:42:45 crc kubenswrapper[4923]: E1009 10:42:45.602600 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:42:57 crc kubenswrapper[4923]: I1009 10:42:57.601795 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:42:57 crc kubenswrapper[4923]: E1009 10:42:57.603780 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:43:09 crc kubenswrapper[4923]: I1009 10:43:09.602357 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:43:09 crc kubenswrapper[4923]: E1009 10:43:09.603640 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.583283 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ng6qh"] Oct 09 10:43:18 crc kubenswrapper[4923]: E1009 10:43:18.584622 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerName="extract-content" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.584645 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerName="extract-content" Oct 09 10:43:18 crc kubenswrapper[4923]: E1009 10:43:18.584659 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerName="extract-utilities" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.584669 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerName="extract-utilities" Oct 09 10:43:18 crc kubenswrapper[4923]: E1009 10:43:18.584686 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerName="extract-utilities" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.584697 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerName="extract-utilities" Oct 09 10:43:18 crc kubenswrapper[4923]: E1009 10:43:18.584724 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerName="registry-server" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.584732 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerName="registry-server" Oct 09 10:43:18 crc kubenswrapper[4923]: E1009 10:43:18.584775 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerName="extract-content" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.584783 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerName="extract-content" Oct 09 10:43:18 crc kubenswrapper[4923]: E1009 10:43:18.584799 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerName="registry-server" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.584807 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerName="registry-server" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.585065 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3e1e21-fe32-403f-b4be-aaadc9379e9d" containerName="registry-server" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.585081 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d70d4cf2-1ab1-4025-864d-37dd1eae658c" containerName="registry-server" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.586846 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.594964 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ng6qh"] Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.657101 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk7zh\" (UniqueName: \"kubernetes.io/projected/7921ebce-2b32-4e08-a849-17cdc9c4c72a-kube-api-access-kk7zh\") pod \"redhat-operators-ng6qh\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.657259 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-catalog-content\") pod \"redhat-operators-ng6qh\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.657372 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-utilities\") pod \"redhat-operators-ng6qh\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.759925 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-utilities\") pod \"redhat-operators-ng6qh\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.760046 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk7zh\" (UniqueName: \"kubernetes.io/projected/7921ebce-2b32-4e08-a849-17cdc9c4c72a-kube-api-access-kk7zh\") pod \"redhat-operators-ng6qh\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.760202 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-catalog-content\") pod \"redhat-operators-ng6qh\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.760790 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-utilities\") pod \"redhat-operators-ng6qh\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.761007 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-catalog-content\") pod \"redhat-operators-ng6qh\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.786036 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk7zh\" (UniqueName: \"kubernetes.io/projected/7921ebce-2b32-4e08-a849-17cdc9c4c72a-kube-api-access-kk7zh\") pod \"redhat-operators-ng6qh\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:18 crc kubenswrapper[4923]: I1009 10:43:18.926850 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:19 crc kubenswrapper[4923]: I1009 10:43:19.417627 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ng6qh"] Oct 09 10:43:19 crc kubenswrapper[4923]: I1009 10:43:19.898136 4923 generic.go:334] "Generic (PLEG): container finished" podID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerID="15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed" exitCode=0 Oct 09 10:43:19 crc kubenswrapper[4923]: I1009 10:43:19.898204 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng6qh" event={"ID":"7921ebce-2b32-4e08-a849-17cdc9c4c72a","Type":"ContainerDied","Data":"15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed"} Oct 09 10:43:19 crc kubenswrapper[4923]: I1009 10:43:19.898243 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng6qh" event={"ID":"7921ebce-2b32-4e08-a849-17cdc9c4c72a","Type":"ContainerStarted","Data":"fd00b78e27bb96aad1efa9994657d55a60905c2d9b41dc7fd463b6c0ba167d87"} Oct 09 10:43:20 crc kubenswrapper[4923]: I1009 10:43:20.602472 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:43:20 crc kubenswrapper[4923]: E1009 10:43:20.603115 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:43:20 crc kubenswrapper[4923]: I1009 10:43:20.946870 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng6qh" event={"ID":"7921ebce-2b32-4e08-a849-17cdc9c4c72a","Type":"ContainerStarted","Data":"a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37"} Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.552475 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jcl9q"] Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.554774 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.570802 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jcl9q"] Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.610264 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16877656-531f-42f1-9c4e-c72e99f6a536-catalog-content\") pod \"certified-operators-jcl9q\" (UID: \"16877656-531f-42f1-9c4e-c72e99f6a536\") " pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.610438 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16877656-531f-42f1-9c4e-c72e99f6a536-utilities\") pod \"certified-operators-jcl9q\" (UID: \"16877656-531f-42f1-9c4e-c72e99f6a536\") " pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.610481 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxcmv\" (UniqueName: \"kubernetes.io/projected/16877656-531f-42f1-9c4e-c72e99f6a536-kube-api-access-hxcmv\") pod \"certified-operators-jcl9q\" (UID: \"16877656-531f-42f1-9c4e-c72e99f6a536\") " pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.712451 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16877656-531f-42f1-9c4e-c72e99f6a536-utilities\") pod \"certified-operators-jcl9q\" (UID: \"16877656-531f-42f1-9c4e-c72e99f6a536\") " pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.712533 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxcmv\" (UniqueName: \"kubernetes.io/projected/16877656-531f-42f1-9c4e-c72e99f6a536-kube-api-access-hxcmv\") pod \"certified-operators-jcl9q\" (UID: \"16877656-531f-42f1-9c4e-c72e99f6a536\") " pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.712609 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16877656-531f-42f1-9c4e-c72e99f6a536-catalog-content\") pod \"certified-operators-jcl9q\" (UID: \"16877656-531f-42f1-9c4e-c72e99f6a536\") " pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.713072 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/16877656-531f-42f1-9c4e-c72e99f6a536-utilities\") pod \"certified-operators-jcl9q\" (UID: \"16877656-531f-42f1-9c4e-c72e99f6a536\") " pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.713088 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/16877656-531f-42f1-9c4e-c72e99f6a536-catalog-content\") pod \"certified-operators-jcl9q\" (UID: \"16877656-531f-42f1-9c4e-c72e99f6a536\") " pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.743547 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxcmv\" (UniqueName: \"kubernetes.io/projected/16877656-531f-42f1-9c4e-c72e99f6a536-kube-api-access-hxcmv\") pod \"certified-operators-jcl9q\" (UID: \"16877656-531f-42f1-9c4e-c72e99f6a536\") " pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.886642 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.959377 4923 generic.go:334] "Generic (PLEG): container finished" podID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerID="a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37" exitCode=0 Oct 09 10:43:21 crc kubenswrapper[4923]: I1009 10:43:21.959443 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng6qh" event={"ID":"7921ebce-2b32-4e08-a849-17cdc9c4c72a","Type":"ContainerDied","Data":"a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37"} Oct 09 10:43:22 crc kubenswrapper[4923]: I1009 10:43:22.498367 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jcl9q"] Oct 09 10:43:22 crc kubenswrapper[4923]: W1009 10:43:22.522148 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16877656_531f_42f1_9c4e_c72e99f6a536.slice/crio-d8596333421fe0970651c8c1b8a98d8c7b9b95dc9020dcfd1742d1fc84be1b8a WatchSource:0}: Error finding container d8596333421fe0970651c8c1b8a98d8c7b9b95dc9020dcfd1742d1fc84be1b8a: Status 404 returned error can't find the container with id d8596333421fe0970651c8c1b8a98d8c7b9b95dc9020dcfd1742d1fc84be1b8a Oct 09 10:43:22 crc kubenswrapper[4923]: I1009 10:43:22.967786 4923 generic.go:334] "Generic (PLEG): container finished" podID="16877656-531f-42f1-9c4e-c72e99f6a536" containerID="9b6270b0de4fa2d2c0cad2a6f275230653ca8d7257194299a26222f657cb0b38" exitCode=0 Oct 09 10:43:22 crc kubenswrapper[4923]: I1009 10:43:22.967851 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jcl9q" event={"ID":"16877656-531f-42f1-9c4e-c72e99f6a536","Type":"ContainerDied","Data":"9b6270b0de4fa2d2c0cad2a6f275230653ca8d7257194299a26222f657cb0b38"} Oct 09 10:43:22 crc kubenswrapper[4923]: I1009 10:43:22.967879 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jcl9q" event={"ID":"16877656-531f-42f1-9c4e-c72e99f6a536","Type":"ContainerStarted","Data":"d8596333421fe0970651c8c1b8a98d8c7b9b95dc9020dcfd1742d1fc84be1b8a"} Oct 09 10:43:22 crc kubenswrapper[4923]: I1009 10:43:22.973464 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng6qh" event={"ID":"7921ebce-2b32-4e08-a849-17cdc9c4c72a","Type":"ContainerStarted","Data":"9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694"} Oct 09 10:43:23 crc kubenswrapper[4923]: I1009 10:43:23.023646 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ng6qh" podStartSLOduration=2.551815017 podStartE2EDuration="5.023625345s" podCreationTimestamp="2025-10-09 10:43:18 +0000 UTC" firstStartedPulling="2025-10-09 10:43:19.900615746 +0000 UTC m=+2285.968797502" lastFinishedPulling="2025-10-09 10:43:22.372426074 +0000 UTC m=+2288.440607830" observedRunningTime="2025-10-09 10:43:23.02233737 +0000 UTC m=+2289.090519126" watchObservedRunningTime="2025-10-09 10:43:23.023625345 +0000 UTC m=+2289.091807101" Oct 09 10:43:28 crc kubenswrapper[4923]: I1009 10:43:28.927651 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:28 crc kubenswrapper[4923]: I1009 10:43:28.929429 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:29 crc kubenswrapper[4923]: I1009 10:43:29.016499 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:29 crc kubenswrapper[4923]: I1009 10:43:29.050092 4923 generic.go:334] "Generic (PLEG): container finished" podID="16877656-531f-42f1-9c4e-c72e99f6a536" containerID="e04e6490cdc710fac838cf76df4e6d16c703ef65deb32ad5d7bebdaaed0955cb" exitCode=0 Oct 09 10:43:29 crc kubenswrapper[4923]: I1009 10:43:29.052879 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jcl9q" event={"ID":"16877656-531f-42f1-9c4e-c72e99f6a536","Type":"ContainerDied","Data":"e04e6490cdc710fac838cf76df4e6d16c703ef65deb32ad5d7bebdaaed0955cb"} Oct 09 10:43:29 crc kubenswrapper[4923]: I1009 10:43:29.104969 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:30 crc kubenswrapper[4923]: I1009 10:43:30.063089 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jcl9q" event={"ID":"16877656-531f-42f1-9c4e-c72e99f6a536","Type":"ContainerStarted","Data":"a9a6d8e193f87509f8dd3c7cad7e5797d56b1547ff721c217a42295ac298c3db"} Oct 09 10:43:30 crc kubenswrapper[4923]: I1009 10:43:30.090729 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jcl9q" podStartSLOduration=2.631402005 podStartE2EDuration="9.090706156s" podCreationTimestamp="2025-10-09 10:43:21 +0000 UTC" firstStartedPulling="2025-10-09 10:43:22.969573877 +0000 UTC m=+2289.037755633" lastFinishedPulling="2025-10-09 10:43:29.428878008 +0000 UTC m=+2295.497059784" observedRunningTime="2025-10-09 10:43:30.081566821 +0000 UTC m=+2296.149748587" watchObservedRunningTime="2025-10-09 10:43:30.090706156 +0000 UTC m=+2296.158887912" Oct 09 10:43:30 crc kubenswrapper[4923]: I1009 10:43:30.265424 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ng6qh"] Oct 09 10:43:31 crc kubenswrapper[4923]: I1009 10:43:31.601973 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:43:31 crc kubenswrapper[4923]: E1009 10:43:31.603003 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:43:31 crc kubenswrapper[4923]: I1009 10:43:31.887655 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:31 crc kubenswrapper[4923]: I1009 10:43:31.887728 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:31 crc kubenswrapper[4923]: I1009 10:43:31.941237 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.080026 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ng6qh" podUID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerName="registry-server" containerID="cri-o://9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694" gracePeriod=2 Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.506607 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.606481 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk7zh\" (UniqueName: \"kubernetes.io/projected/7921ebce-2b32-4e08-a849-17cdc9c4c72a-kube-api-access-kk7zh\") pod \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.606528 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-utilities\") pod \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.606636 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-catalog-content\") pod \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\" (UID: \"7921ebce-2b32-4e08-a849-17cdc9c4c72a\") " Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.607087 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-utilities" (OuterVolumeSpecName: "utilities") pod "7921ebce-2b32-4e08-a849-17cdc9c4c72a" (UID: "7921ebce-2b32-4e08-a849-17cdc9c4c72a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.611904 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7921ebce-2b32-4e08-a849-17cdc9c4c72a-kube-api-access-kk7zh" (OuterVolumeSpecName: "kube-api-access-kk7zh") pod "7921ebce-2b32-4e08-a849-17cdc9c4c72a" (UID: "7921ebce-2b32-4e08-a849-17cdc9c4c72a"). InnerVolumeSpecName "kube-api-access-kk7zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.687091 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7921ebce-2b32-4e08-a849-17cdc9c4c72a" (UID: "7921ebce-2b32-4e08-a849-17cdc9c4c72a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.710680 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk7zh\" (UniqueName: \"kubernetes.io/projected/7921ebce-2b32-4e08-a849-17cdc9c4c72a-kube-api-access-kk7zh\") on node \"crc\" DevicePath \"\"" Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.710717 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:43:32 crc kubenswrapper[4923]: I1009 10:43:32.710726 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7921ebce-2b32-4e08-a849-17cdc9c4c72a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.093674 4923 generic.go:334] "Generic (PLEG): container finished" podID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerID="9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694" exitCode=0 Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.093747 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng6qh" event={"ID":"7921ebce-2b32-4e08-a849-17cdc9c4c72a","Type":"ContainerDied","Data":"9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694"} Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.093824 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ng6qh" event={"ID":"7921ebce-2b32-4e08-a849-17cdc9c4c72a","Type":"ContainerDied","Data":"fd00b78e27bb96aad1efa9994657d55a60905c2d9b41dc7fd463b6c0ba167d87"} Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.093830 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ng6qh" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.093846 4923 scope.go:117] "RemoveContainer" containerID="9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.124383 4923 scope.go:117] "RemoveContainer" containerID="a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.130881 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ng6qh"] Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.138705 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ng6qh"] Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.150371 4923 scope.go:117] "RemoveContainer" containerID="15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.226312 4923 scope.go:117] "RemoveContainer" containerID="9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694" Oct 09 10:43:33 crc kubenswrapper[4923]: E1009 10:43:33.226940 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694\": container with ID starting with 9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694 not found: ID does not exist" containerID="9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.226994 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694"} err="failed to get container status \"9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694\": rpc error: code = NotFound desc = could not find container \"9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694\": container with ID starting with 9c15c8fd77afccacd754edef796fe7ad2539cb9d082e8bcf3194d4602af8f694 not found: ID does not exist" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.227030 4923 scope.go:117] "RemoveContainer" containerID="a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37" Oct 09 10:43:33 crc kubenswrapper[4923]: E1009 10:43:33.227662 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37\": container with ID starting with a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37 not found: ID does not exist" containerID="a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.227745 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37"} err="failed to get container status \"a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37\": rpc error: code = NotFound desc = could not find container \"a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37\": container with ID starting with a4db51830423a6e75d703c38268d73d8ca332ecc12a5a09c8c62e2ed29c9ce37 not found: ID does not exist" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.227842 4923 scope.go:117] "RemoveContainer" containerID="15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed" Oct 09 10:43:33 crc kubenswrapper[4923]: E1009 10:43:33.228279 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed\": container with ID starting with 15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed not found: ID does not exist" containerID="15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed" Oct 09 10:43:33 crc kubenswrapper[4923]: I1009 10:43:33.228310 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed"} err="failed to get container status \"15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed\": rpc error: code = NotFound desc = could not find container \"15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed\": container with ID starting with 15e2d807b323cb27dfbdbe2c5eadcbf9803380b919738daefd0bc8457d6298ed not found: ID does not exist" Oct 09 10:43:34 crc kubenswrapper[4923]: I1009 10:43:34.613735 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" path="/var/lib/kubelet/pods/7921ebce-2b32-4e08-a849-17cdc9c4c72a/volumes" Oct 09 10:43:41 crc kubenswrapper[4923]: I1009 10:43:41.995205 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jcl9q" Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.138565 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jcl9q"] Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.203987 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hbbk5"] Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.204768 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hbbk5" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerName="registry-server" containerID="cri-o://112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4" gracePeriod=2 Oct 09 10:43:42 crc kubenswrapper[4923]: E1009 10:43:42.473299 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4 is running failed: container process not found" containerID="112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4" cmd=["grpc_health_probe","-addr=:50051"] Oct 09 10:43:42 crc kubenswrapper[4923]: E1009 10:43:42.473940 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4 is running failed: container process not found" containerID="112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4" cmd=["grpc_health_probe","-addr=:50051"] Oct 09 10:43:42 crc kubenswrapper[4923]: E1009 10:43:42.474404 4923 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4 is running failed: container process not found" containerID="112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4" cmd=["grpc_health_probe","-addr=:50051"] Oct 09 10:43:42 crc kubenswrapper[4923]: E1009 10:43:42.474457 4923 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-hbbk5" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerName="registry-server" Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.778867 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.864562 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75cdc\" (UniqueName: \"kubernetes.io/projected/09837ba0-2b96-4f41-bb93-268a05cf3f2d-kube-api-access-75cdc\") pod \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.865028 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-catalog-content\") pod \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.865222 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-utilities\") pod \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\" (UID: \"09837ba0-2b96-4f41-bb93-268a05cf3f2d\") " Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.865972 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-utilities" (OuterVolumeSpecName: "utilities") pod "09837ba0-2b96-4f41-bb93-268a05cf3f2d" (UID: "09837ba0-2b96-4f41-bb93-268a05cf3f2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.866728 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.874825 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09837ba0-2b96-4f41-bb93-268a05cf3f2d-kube-api-access-75cdc" (OuterVolumeSpecName: "kube-api-access-75cdc") pod "09837ba0-2b96-4f41-bb93-268a05cf3f2d" (UID: "09837ba0-2b96-4f41-bb93-268a05cf3f2d"). InnerVolumeSpecName "kube-api-access-75cdc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.920024 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09837ba0-2b96-4f41-bb93-268a05cf3f2d" (UID: "09837ba0-2b96-4f41-bb93-268a05cf3f2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.968716 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75cdc\" (UniqueName: \"kubernetes.io/projected/09837ba0-2b96-4f41-bb93-268a05cf3f2d-kube-api-access-75cdc\") on node \"crc\" DevicePath \"\"" Oct 09 10:43:42 crc kubenswrapper[4923]: I1009 10:43:42.969026 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09837ba0-2b96-4f41-bb93-268a05cf3f2d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.207319 4923 generic.go:334] "Generic (PLEG): container finished" podID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerID="112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4" exitCode=0 Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.207381 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hbbk5" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.207389 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbbk5" event={"ID":"09837ba0-2b96-4f41-bb93-268a05cf3f2d","Type":"ContainerDied","Data":"112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4"} Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.207503 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hbbk5" event={"ID":"09837ba0-2b96-4f41-bb93-268a05cf3f2d","Type":"ContainerDied","Data":"f47b11e9b70e23234c6946e6d161607f4875ccddf5c25d95161e8695c85383c4"} Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.207537 4923 scope.go:117] "RemoveContainer" containerID="112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.241585 4923 scope.go:117] "RemoveContainer" containerID="55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.248917 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hbbk5"] Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.260769 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hbbk5"] Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.269556 4923 scope.go:117] "RemoveContainer" containerID="a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.350018 4923 scope.go:117] "RemoveContainer" containerID="112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4" Oct 09 10:43:43 crc kubenswrapper[4923]: E1009 10:43:43.350633 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4\": container with ID starting with 112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4 not found: ID does not exist" containerID="112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.350669 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4"} err="failed to get container status \"112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4\": rpc error: code = NotFound desc = could not find container \"112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4\": container with ID starting with 112f5ebed610bca15b14c31b79e0d6d5c9b10f41bdb3beb518f55692f73c18f4 not found: ID does not exist" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.350698 4923 scope.go:117] "RemoveContainer" containerID="55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0" Oct 09 10:43:43 crc kubenswrapper[4923]: E1009 10:43:43.351568 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0\": container with ID starting with 55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0 not found: ID does not exist" containerID="55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.351602 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0"} err="failed to get container status \"55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0\": rpc error: code = NotFound desc = could not find container \"55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0\": container with ID starting with 55125f1b3a536b2d169da75fb37bb0d673913a1fa364867a4966f396be027bc0 not found: ID does not exist" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.351624 4923 scope.go:117] "RemoveContainer" containerID="a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd" Oct 09 10:43:43 crc kubenswrapper[4923]: E1009 10:43:43.351979 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd\": container with ID starting with a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd not found: ID does not exist" containerID="a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd" Oct 09 10:43:43 crc kubenswrapper[4923]: I1009 10:43:43.352004 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd"} err="failed to get container status \"a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd\": rpc error: code = NotFound desc = could not find container \"a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd\": container with ID starting with a17cdd6d2040b73452de5bea3c23ce16aac978f2547e52802c3619828aff19fd not found: ID does not exist" Oct 09 10:43:44 crc kubenswrapper[4923]: I1009 10:43:44.613594 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:43:44 crc kubenswrapper[4923]: E1009 10:43:44.616936 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:43:44 crc kubenswrapper[4923]: I1009 10:43:44.636784 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" path="/var/lib/kubelet/pods/09837ba0-2b96-4f41-bb93-268a05cf3f2d/volumes" Oct 09 10:43:58 crc kubenswrapper[4923]: I1009 10:43:58.602569 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:43:58 crc kubenswrapper[4923]: E1009 10:43:58.604577 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.024815 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.040004 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-4wfvf"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.051147 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.058042 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.063768 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.069111 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w4cbw"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.078660 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.085680 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.090955 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.096722 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-njnq8"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.101955 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w4cbw"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.107719 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqtjj"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.115245 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.123546 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.130181 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-sfcpb"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.135872 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pnm2b"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.144065 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-xpm6t"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.150263 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-2ckb7"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.155956 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-2279h"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.161469 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-slbn6"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.166817 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5"] Oct 09 10:44:09 crc kubenswrapper[4923]: I1009 10:44:09.172157 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-q7pb5"] Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.602142 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:44:10 crc kubenswrapper[4923]: E1009 10:44:10.602992 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.615910 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00251bc8-62c6-4f4e-8aa0-f8da07570f3a" path="/var/lib/kubelet/pods/00251bc8-62c6-4f4e-8aa0-f8da07570f3a/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.616829 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0db9879d-8e0a-4d29-9168-caf8e16d5c3d" path="/var/lib/kubelet/pods/0db9879d-8e0a-4d29-9168-caf8e16d5c3d/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.617599 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d78a009-a62b-41e8-8384-3c1f668ced70" path="/var/lib/kubelet/pods/1d78a009-a62b-41e8-8384-3c1f668ced70/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.618433 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d901fa9-21b4-4b52-bbd1-11aa7cd06e58" path="/var/lib/kubelet/pods/1d901fa9-21b4-4b52-bbd1-11aa7cd06e58/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.619992 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d865110-029b-4108-8e59-cb3947a6269b" path="/var/lib/kubelet/pods/2d865110-029b-4108-8e59-cb3947a6269b/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.620827 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69707bda-a6a2-4e7f-98c4-63a65c9d26fb" path="/var/lib/kubelet/pods/69707bda-a6a2-4e7f-98c4-63a65c9d26fb/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.621641 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ba3fa71-9b42-4f4c-a1de-297d81d72269" path="/var/lib/kubelet/pods/9ba3fa71-9b42-4f4c-a1de-297d81d72269/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.623623 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f5ca5ba-bd14-40f8-aab3-094e1771a374" path="/var/lib/kubelet/pods/9f5ca5ba-bd14-40f8-aab3-094e1771a374/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.624958 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc3626c2-ad33-40b5-a56d-1e4ea07696f2" path="/var/lib/kubelet/pods/cc3626c2-ad33-40b5-a56d-1e4ea07696f2/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.625774 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d54c1adc-c247-4cd6-a97d-843ea434cad3" path="/var/lib/kubelet/pods/d54c1adc-c247-4cd6-a97d-843ea434cad3/volumes" Oct 09 10:44:10 crc kubenswrapper[4923]: I1009 10:44:10.627672 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff2e838f-c4ce-43db-bb07-4cd0191d0079" path="/var/lib/kubelet/pods/ff2e838f-c4ce-43db-bb07-4cd0191d0079/volumes" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.474112 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88"] Oct 09 10:44:14 crc kubenswrapper[4923]: E1009 10:44:14.475131 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerName="extract-utilities" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.475153 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerName="extract-utilities" Oct 09 10:44:14 crc kubenswrapper[4923]: E1009 10:44:14.475168 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerName="extract-utilities" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.475177 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerName="extract-utilities" Oct 09 10:44:14 crc kubenswrapper[4923]: E1009 10:44:14.475196 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerName="registry-server" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.475205 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerName="registry-server" Oct 09 10:44:14 crc kubenswrapper[4923]: E1009 10:44:14.475223 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerName="registry-server" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.475232 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerName="registry-server" Oct 09 10:44:14 crc kubenswrapper[4923]: E1009 10:44:14.475260 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerName="extract-content" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.475268 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerName="extract-content" Oct 09 10:44:14 crc kubenswrapper[4923]: E1009 10:44:14.475294 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerName="extract-content" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.475302 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerName="extract-content" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.475555 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="09837ba0-2b96-4f41-bb93-268a05cf3f2d" containerName="registry-server" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.475581 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="7921ebce-2b32-4e08-a849-17cdc9c4c72a" containerName="registry-server" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.476416 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.479215 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.479848 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.480122 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.480520 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.482308 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.487257 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.487296 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.487327 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.487530 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.487577 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dcj6\" (UniqueName: \"kubernetes.io/projected/9921186f-446f-485d-b39b-e86be6df0511-kube-api-access-6dcj6\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.491715 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88"] Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.593473 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.593824 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dcj6\" (UniqueName: \"kubernetes.io/projected/9921186f-446f-485d-b39b-e86be6df0511-kube-api-access-6dcj6\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.594015 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.594134 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.594235 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.599303 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.599925 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.600151 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.602465 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.613847 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.616162 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.616348 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.621746 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dcj6\" (UniqueName: \"kubernetes.io/projected/9921186f-446f-485d-b39b-e86be6df0511-kube-api-access-6dcj6\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.801228 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:44:14 crc kubenswrapper[4923]: I1009 10:44:14.810301 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:15 crc kubenswrapper[4923]: I1009 10:44:15.339742 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88"] Oct 09 10:44:15 crc kubenswrapper[4923]: W1009 10:44:15.342953 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9921186f_446f_485d_b39b_e86be6df0511.slice/crio-1f2da916c3a3c4c7a7fc9be082dc4c14378d00697fa92c6ef8391031439b65a6 WatchSource:0}: Error finding container 1f2da916c3a3c4c7a7fc9be082dc4c14378d00697fa92c6ef8391031439b65a6: Status 404 returned error can't find the container with id 1f2da916c3a3c4c7a7fc9be082dc4c14378d00697fa92c6ef8391031439b65a6 Oct 09 10:44:15 crc kubenswrapper[4923]: I1009 10:44:15.518460 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" event={"ID":"9921186f-446f-485d-b39b-e86be6df0511","Type":"ContainerStarted","Data":"1f2da916c3a3c4c7a7fc9be082dc4c14378d00697fa92c6ef8391031439b65a6"} Oct 09 10:44:15 crc kubenswrapper[4923]: I1009 10:44:15.764543 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:44:16 crc kubenswrapper[4923]: I1009 10:44:16.529043 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" event={"ID":"9921186f-446f-485d-b39b-e86be6df0511","Type":"ContainerStarted","Data":"ffc3e7a49f7804b1f52a71a28311bd44186a95ff7bd5262961811206916bfc48"} Oct 09 10:44:16 crc kubenswrapper[4923]: I1009 10:44:16.557178 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" podStartSLOduration=2.141114367 podStartE2EDuration="2.557120395s" podCreationTimestamp="2025-10-09 10:44:14 +0000 UTC" firstStartedPulling="2025-10-09 10:44:15.345535611 +0000 UTC m=+2341.413717367" lastFinishedPulling="2025-10-09 10:44:15.761541639 +0000 UTC m=+2341.829723395" observedRunningTime="2025-10-09 10:44:16.556072206 +0000 UTC m=+2342.624254012" watchObservedRunningTime="2025-10-09 10:44:16.557120395 +0000 UTC m=+2342.625302161" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.161075 4923 scope.go:117] "RemoveContainer" containerID="7513b828ccc587c4bd7f26721212b18413bd6398bbaae3f9f439ab9c1b356349" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.209370 4923 scope.go:117] "RemoveContainer" containerID="5bb423e27afece4cbcc63d052ac9dd55f83ac8f57afea95e5297dbd11df9257e" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.284003 4923 scope.go:117] "RemoveContainer" containerID="6f0cd671f826c9d29056422776a0cf628ae70499aefc6b58dccc8b71acb825ea" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.323654 4923 scope.go:117] "RemoveContainer" containerID="29aac3997105a778d531276d0b53f750c56e70f4235ed3bafa5a15890e43455a" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.371424 4923 scope.go:117] "RemoveContainer" containerID="40c4a05b7499f8c88b02b29ec6017cf46657afe8f377dbd3995a2317df9b2297" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.427609 4923 scope.go:117] "RemoveContainer" containerID="6431403dfee5423036e307a91cd66a480d43d6359485ccedc4cba1d163f7b4cc" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.478623 4923 scope.go:117] "RemoveContainer" containerID="52fa1ed0b40d689b747beaa7ce3213bf7ce43d82ed8ac44e9c75139f27c9891e" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.514294 4923 scope.go:117] "RemoveContainer" containerID="51ca72b4cdd99129b1f942fa19d543f88fb5745bedc7905827dbf9dfd25a2f6b" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.544216 4923 scope.go:117] "RemoveContainer" containerID="c6e06ea5c1a11c5a00add805c5240fdd44f274e5f7bd9baf666ba0b71e725e67" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.586167 4923 scope.go:117] "RemoveContainer" containerID="ec8e791db8671720b772fd554a378ba2dcce36ecdfb5d4e7343d97abbcb11535" Oct 09 10:44:19 crc kubenswrapper[4923]: I1009 10:44:19.632299 4923 scope.go:117] "RemoveContainer" containerID="4f2c0b6cf0546439a2a9fc4a64379397479861e1adeec4afff360b6186dc6fa5" Oct 09 10:44:23 crc kubenswrapper[4923]: I1009 10:44:23.601315 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:44:23 crc kubenswrapper[4923]: E1009 10:44:23.602326 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:44:27 crc kubenswrapper[4923]: I1009 10:44:27.663951 4923 generic.go:334] "Generic (PLEG): container finished" podID="9921186f-446f-485d-b39b-e86be6df0511" containerID="ffc3e7a49f7804b1f52a71a28311bd44186a95ff7bd5262961811206916bfc48" exitCode=0 Oct 09 10:44:27 crc kubenswrapper[4923]: I1009 10:44:27.664056 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" event={"ID":"9921186f-446f-485d-b39b-e86be6df0511","Type":"ContainerDied","Data":"ffc3e7a49f7804b1f52a71a28311bd44186a95ff7bd5262961811206916bfc48"} Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.085577 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.136830 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-inventory\") pod \"9921186f-446f-485d-b39b-e86be6df0511\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.136899 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-repo-setup-combined-ca-bundle\") pod \"9921186f-446f-485d-b39b-e86be6df0511\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.136961 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ceph\") pod \"9921186f-446f-485d-b39b-e86be6df0511\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.137064 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ssh-key\") pod \"9921186f-446f-485d-b39b-e86be6df0511\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.137184 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dcj6\" (UniqueName: \"kubernetes.io/projected/9921186f-446f-485d-b39b-e86be6df0511-kube-api-access-6dcj6\") pod \"9921186f-446f-485d-b39b-e86be6df0511\" (UID: \"9921186f-446f-485d-b39b-e86be6df0511\") " Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.147012 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9921186f-446f-485d-b39b-e86be6df0511-kube-api-access-6dcj6" (OuterVolumeSpecName: "kube-api-access-6dcj6") pod "9921186f-446f-485d-b39b-e86be6df0511" (UID: "9921186f-446f-485d-b39b-e86be6df0511"). InnerVolumeSpecName "kube-api-access-6dcj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.147143 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ceph" (OuterVolumeSpecName: "ceph") pod "9921186f-446f-485d-b39b-e86be6df0511" (UID: "9921186f-446f-485d-b39b-e86be6df0511"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.148833 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "9921186f-446f-485d-b39b-e86be6df0511" (UID: "9921186f-446f-485d-b39b-e86be6df0511"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.172174 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-inventory" (OuterVolumeSpecName: "inventory") pod "9921186f-446f-485d-b39b-e86be6df0511" (UID: "9921186f-446f-485d-b39b-e86be6df0511"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.180977 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9921186f-446f-485d-b39b-e86be6df0511" (UID: "9921186f-446f-485d-b39b-e86be6df0511"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.238638 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.238678 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dcj6\" (UniqueName: \"kubernetes.io/projected/9921186f-446f-485d-b39b-e86be6df0511-kube-api-access-6dcj6\") on node \"crc\" DevicePath \"\"" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.238694 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.238709 4923 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.238721 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9921186f-446f-485d-b39b-e86be6df0511-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.684080 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" event={"ID":"9921186f-446f-485d-b39b-e86be6df0511","Type":"ContainerDied","Data":"1f2da916c3a3c4c7a7fc9be082dc4c14378d00697fa92c6ef8391031439b65a6"} Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.684156 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f2da916c3a3c4c7a7fc9be082dc4c14378d00697fa92c6ef8391031439b65a6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.684164 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.774750 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6"] Oct 09 10:44:29 crc kubenswrapper[4923]: E1009 10:44:29.775227 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9921186f-446f-485d-b39b-e86be6df0511" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.775250 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9921186f-446f-485d-b39b-e86be6df0511" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.775563 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9921186f-446f-485d-b39b-e86be6df0511" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.776341 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.780593 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.781113 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.781110 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.781167 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.781171 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.804244 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6"] Oct 09 10:44:29 crc kubenswrapper[4923]: E1009 10:44:29.841682 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9921186f_446f_485d_b39b_e86be6df0511.slice\": RecentStats: unable to find data in memory cache]" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.852482 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.852545 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.852575 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.852610 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.852656 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz5bp\" (UniqueName: \"kubernetes.io/projected/3f45569e-63f8-4693-9b2c-6b6e8d75067b-kube-api-access-jz5bp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.955468 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.955632 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.955713 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.955808 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.955904 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz5bp\" (UniqueName: \"kubernetes.io/projected/3f45569e-63f8-4693-9b2c-6b6e8d75067b-kube-api-access-jz5bp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.963507 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.963525 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.964149 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.970343 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:29 crc kubenswrapper[4923]: I1009 10:44:29.973324 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz5bp\" (UniqueName: \"kubernetes.io/projected/3f45569e-63f8-4693-9b2c-6b6e8d75067b-kube-api-access-jz5bp\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:30 crc kubenswrapper[4923]: I1009 10:44:30.116287 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:44:30 crc kubenswrapper[4923]: I1009 10:44:30.640763 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6"] Oct 09 10:44:30 crc kubenswrapper[4923]: W1009 10:44:30.644076 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f45569e_63f8_4693_9b2c_6b6e8d75067b.slice/crio-8d8e11096e3b0c12d39619bc6a51173a049567de644862cd9e6c15f6eb90ab88 WatchSource:0}: Error finding container 8d8e11096e3b0c12d39619bc6a51173a049567de644862cd9e6c15f6eb90ab88: Status 404 returned error can't find the container with id 8d8e11096e3b0c12d39619bc6a51173a049567de644862cd9e6c15f6eb90ab88 Oct 09 10:44:30 crc kubenswrapper[4923]: I1009 10:44:30.696243 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" event={"ID":"3f45569e-63f8-4693-9b2c-6b6e8d75067b","Type":"ContainerStarted","Data":"8d8e11096e3b0c12d39619bc6a51173a049567de644862cd9e6c15f6eb90ab88"} Oct 09 10:44:31 crc kubenswrapper[4923]: I1009 10:44:31.724986 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" event={"ID":"3f45569e-63f8-4693-9b2c-6b6e8d75067b","Type":"ContainerStarted","Data":"2b45c98c2b0a0a840cb90292ebff76b03c9c68c79b966f58834d14fd08a5a8b1"} Oct 09 10:44:31 crc kubenswrapper[4923]: I1009 10:44:31.751843 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" podStartSLOduration=2.256240367 podStartE2EDuration="2.751812532s" podCreationTimestamp="2025-10-09 10:44:29 +0000 UTC" firstStartedPulling="2025-10-09 10:44:30.64630816 +0000 UTC m=+2356.714489906" lastFinishedPulling="2025-10-09 10:44:31.141880305 +0000 UTC m=+2357.210062071" observedRunningTime="2025-10-09 10:44:31.746868095 +0000 UTC m=+2357.815049851" watchObservedRunningTime="2025-10-09 10:44:31.751812532 +0000 UTC m=+2357.819994298" Oct 09 10:44:37 crc kubenswrapper[4923]: I1009 10:44:37.601793 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:44:37 crc kubenswrapper[4923]: E1009 10:44:37.603002 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:44:49 crc kubenswrapper[4923]: I1009 10:44:49.601546 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:44:49 crc kubenswrapper[4923]: E1009 10:44:49.602598 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.152674 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm"] Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.155319 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.159122 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.159208 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.179993 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm"] Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.266123 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05baf8e1-8781-427d-aaf9-f01b3b744509-config-volume\") pod \"collect-profiles-29333445-8gflm\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.266185 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05baf8e1-8781-427d-aaf9-f01b3b744509-secret-volume\") pod \"collect-profiles-29333445-8gflm\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.266242 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6cdk\" (UniqueName: \"kubernetes.io/projected/05baf8e1-8781-427d-aaf9-f01b3b744509-kube-api-access-v6cdk\") pod \"collect-profiles-29333445-8gflm\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.368923 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05baf8e1-8781-427d-aaf9-f01b3b744509-config-volume\") pod \"collect-profiles-29333445-8gflm\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.369008 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05baf8e1-8781-427d-aaf9-f01b3b744509-secret-volume\") pod \"collect-profiles-29333445-8gflm\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.369047 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6cdk\" (UniqueName: \"kubernetes.io/projected/05baf8e1-8781-427d-aaf9-f01b3b744509-kube-api-access-v6cdk\") pod \"collect-profiles-29333445-8gflm\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.370453 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05baf8e1-8781-427d-aaf9-f01b3b744509-config-volume\") pod \"collect-profiles-29333445-8gflm\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.381448 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05baf8e1-8781-427d-aaf9-f01b3b744509-secret-volume\") pod \"collect-profiles-29333445-8gflm\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.389979 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6cdk\" (UniqueName: \"kubernetes.io/projected/05baf8e1-8781-427d-aaf9-f01b3b744509-kube-api-access-v6cdk\") pod \"collect-profiles-29333445-8gflm\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.478955 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.602064 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:45:00 crc kubenswrapper[4923]: E1009 10:45:00.602596 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:45:00 crc kubenswrapper[4923]: I1009 10:45:00.949262 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm"] Oct 09 10:45:01 crc kubenswrapper[4923]: I1009 10:45:01.002580 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" event={"ID":"05baf8e1-8781-427d-aaf9-f01b3b744509","Type":"ContainerStarted","Data":"22c1542190833dcc705652c5c2c1bc8c3e11c5fc8e280dd427f6c2e5f3b15155"} Oct 09 10:45:02 crc kubenswrapper[4923]: I1009 10:45:02.012016 4923 generic.go:334] "Generic (PLEG): container finished" podID="05baf8e1-8781-427d-aaf9-f01b3b744509" containerID="c1e2aeaee21aa9260d4b24f920f8e5c49acfb884787b1e3c07f36f1f6b2d0fb9" exitCode=0 Oct 09 10:45:02 crc kubenswrapper[4923]: I1009 10:45:02.012194 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" event={"ID":"05baf8e1-8781-427d-aaf9-f01b3b744509","Type":"ContainerDied","Data":"c1e2aeaee21aa9260d4b24f920f8e5c49acfb884787b1e3c07f36f1f6b2d0fb9"} Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.356033 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.536821 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05baf8e1-8781-427d-aaf9-f01b3b744509-config-volume\") pod \"05baf8e1-8781-427d-aaf9-f01b3b744509\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.537222 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05baf8e1-8781-427d-aaf9-f01b3b744509-secret-volume\") pod \"05baf8e1-8781-427d-aaf9-f01b3b744509\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.537418 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6cdk\" (UniqueName: \"kubernetes.io/projected/05baf8e1-8781-427d-aaf9-f01b3b744509-kube-api-access-v6cdk\") pod \"05baf8e1-8781-427d-aaf9-f01b3b744509\" (UID: \"05baf8e1-8781-427d-aaf9-f01b3b744509\") " Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.537870 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05baf8e1-8781-427d-aaf9-f01b3b744509-config-volume" (OuterVolumeSpecName: "config-volume") pod "05baf8e1-8781-427d-aaf9-f01b3b744509" (UID: "05baf8e1-8781-427d-aaf9-f01b3b744509"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.538237 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/05baf8e1-8781-427d-aaf9-f01b3b744509-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.543034 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05baf8e1-8781-427d-aaf9-f01b3b744509-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "05baf8e1-8781-427d-aaf9-f01b3b744509" (UID: "05baf8e1-8781-427d-aaf9-f01b3b744509"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.543361 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05baf8e1-8781-427d-aaf9-f01b3b744509-kube-api-access-v6cdk" (OuterVolumeSpecName: "kube-api-access-v6cdk") pod "05baf8e1-8781-427d-aaf9-f01b3b744509" (UID: "05baf8e1-8781-427d-aaf9-f01b3b744509"). InnerVolumeSpecName "kube-api-access-v6cdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.640894 4923 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/05baf8e1-8781-427d-aaf9-f01b3b744509-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 10:45:03 crc kubenswrapper[4923]: I1009 10:45:03.640945 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6cdk\" (UniqueName: \"kubernetes.io/projected/05baf8e1-8781-427d-aaf9-f01b3b744509-kube-api-access-v6cdk\") on node \"crc\" DevicePath \"\"" Oct 09 10:45:04 crc kubenswrapper[4923]: I1009 10:45:04.031935 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" event={"ID":"05baf8e1-8781-427d-aaf9-f01b3b744509","Type":"ContainerDied","Data":"22c1542190833dcc705652c5c2c1bc8c3e11c5fc8e280dd427f6c2e5f3b15155"} Oct 09 10:45:04 crc kubenswrapper[4923]: I1009 10:45:04.032009 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22c1542190833dcc705652c5c2c1bc8c3e11c5fc8e280dd427f6c2e5f3b15155" Oct 09 10:45:04 crc kubenswrapper[4923]: I1009 10:45:04.032013 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333445-8gflm" Oct 09 10:45:04 crc kubenswrapper[4923]: I1009 10:45:04.430594 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2"] Oct 09 10:45:04 crc kubenswrapper[4923]: I1009 10:45:04.438144 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333400-d47m2"] Oct 09 10:45:04 crc kubenswrapper[4923]: I1009 10:45:04.648045 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="215543fc-99e1-47f3-9ffd-9adb4fd9c611" path="/var/lib/kubelet/pods/215543fc-99e1-47f3-9ffd-9adb4fd9c611/volumes" Oct 09 10:45:14 crc kubenswrapper[4923]: I1009 10:45:14.613872 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:45:14 crc kubenswrapper[4923]: E1009 10:45:14.616564 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:45:19 crc kubenswrapper[4923]: I1009 10:45:19.874461 4923 scope.go:117] "RemoveContainer" containerID="d494dc0a2d9cf9f20441b740e4d1a3fb51473fb0a9ed890bfb6b0da48032c79b" Oct 09 10:45:26 crc kubenswrapper[4923]: I1009 10:45:26.603175 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:45:26 crc kubenswrapper[4923]: E1009 10:45:26.603969 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:45:41 crc kubenswrapper[4923]: I1009 10:45:41.602427 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:45:41 crc kubenswrapper[4923]: E1009 10:45:41.604085 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:45:56 crc kubenswrapper[4923]: I1009 10:45:56.603101 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:45:56 crc kubenswrapper[4923]: E1009 10:45:56.604160 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:46:04 crc kubenswrapper[4923]: I1009 10:46:04.602849 4923 generic.go:334] "Generic (PLEG): container finished" podID="3f45569e-63f8-4693-9b2c-6b6e8d75067b" containerID="2b45c98c2b0a0a840cb90292ebff76b03c9c68c79b966f58834d14fd08a5a8b1" exitCode=0 Oct 09 10:46:04 crc kubenswrapper[4923]: I1009 10:46:04.612681 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" event={"ID":"3f45569e-63f8-4693-9b2c-6b6e8d75067b","Type":"ContainerDied","Data":"2b45c98c2b0a0a840cb90292ebff76b03c9c68c79b966f58834d14fd08a5a8b1"} Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.050324 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.170094 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-bootstrap-combined-ca-bundle\") pod \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.170152 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ceph\") pod \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.170463 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz5bp\" (UniqueName: \"kubernetes.io/projected/3f45569e-63f8-4693-9b2c-6b6e8d75067b-kube-api-access-jz5bp\") pod \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.170537 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ssh-key\") pod \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.170605 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-inventory\") pod \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\" (UID: \"3f45569e-63f8-4693-9b2c-6b6e8d75067b\") " Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.177031 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "3f45569e-63f8-4693-9b2c-6b6e8d75067b" (UID: "3f45569e-63f8-4693-9b2c-6b6e8d75067b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.186791 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ceph" (OuterVolumeSpecName: "ceph") pod "3f45569e-63f8-4693-9b2c-6b6e8d75067b" (UID: "3f45569e-63f8-4693-9b2c-6b6e8d75067b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.189512 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f45569e-63f8-4693-9b2c-6b6e8d75067b-kube-api-access-jz5bp" (OuterVolumeSpecName: "kube-api-access-jz5bp") pod "3f45569e-63f8-4693-9b2c-6b6e8d75067b" (UID: "3f45569e-63f8-4693-9b2c-6b6e8d75067b"). InnerVolumeSpecName "kube-api-access-jz5bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.200905 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-inventory" (OuterVolumeSpecName: "inventory") pod "3f45569e-63f8-4693-9b2c-6b6e8d75067b" (UID: "3f45569e-63f8-4693-9b2c-6b6e8d75067b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.205893 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3f45569e-63f8-4693-9b2c-6b6e8d75067b" (UID: "3f45569e-63f8-4693-9b2c-6b6e8d75067b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.273425 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz5bp\" (UniqueName: \"kubernetes.io/projected/3f45569e-63f8-4693-9b2c-6b6e8d75067b-kube-api-access-jz5bp\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.273460 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.273473 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.273482 4923 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.273493 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3f45569e-63f8-4693-9b2c-6b6e8d75067b-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.624742 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" event={"ID":"3f45569e-63f8-4693-9b2c-6b6e8d75067b","Type":"ContainerDied","Data":"8d8e11096e3b0c12d39619bc6a51173a049567de644862cd9e6c15f6eb90ab88"} Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.624827 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.624836 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d8e11096e3b0c12d39619bc6a51173a049567de644862cd9e6c15f6eb90ab88" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.732627 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w"] Oct 09 10:46:06 crc kubenswrapper[4923]: E1009 10:46:06.733092 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05baf8e1-8781-427d-aaf9-f01b3b744509" containerName="collect-profiles" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.733118 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="05baf8e1-8781-427d-aaf9-f01b3b744509" containerName="collect-profiles" Oct 09 10:46:06 crc kubenswrapper[4923]: E1009 10:46:06.733140 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f45569e-63f8-4693-9b2c-6b6e8d75067b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.733151 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f45569e-63f8-4693-9b2c-6b6e8d75067b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.733390 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="05baf8e1-8781-427d-aaf9-f01b3b744509" containerName="collect-profiles" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.733423 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f45569e-63f8-4693-9b2c-6b6e8d75067b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.734287 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.739275 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.739282 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.739290 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.739665 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.739884 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.752073 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w"] Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.784090 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xbft\" (UniqueName: \"kubernetes.io/projected/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-kube-api-access-4xbft\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.784303 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.784347 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.784430 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.886444 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.886503 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.886593 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.886678 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xbft\" (UniqueName: \"kubernetes.io/projected/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-kube-api-access-4xbft\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.890335 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.890374 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.894474 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:06 crc kubenswrapper[4923]: I1009 10:46:06.906338 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xbft\" (UniqueName: \"kubernetes.io/projected/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-kube-api-access-4xbft\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:07 crc kubenswrapper[4923]: I1009 10:46:07.053056 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:07 crc kubenswrapper[4923]: I1009 10:46:07.617304 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w"] Oct 09 10:46:07 crc kubenswrapper[4923]: I1009 10:46:07.640704 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 10:46:08 crc kubenswrapper[4923]: I1009 10:46:08.671229 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" event={"ID":"d8452ab6-f295-464d-bf20-c4a0d7f6ce64","Type":"ContainerStarted","Data":"002d3a92aa9a8de6d170ca2c674335c8ad7bfc1ee2668bbd2fe45216aeb57f16"} Oct 09 10:46:08 crc kubenswrapper[4923]: I1009 10:46:08.671694 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" event={"ID":"d8452ab6-f295-464d-bf20-c4a0d7f6ce64","Type":"ContainerStarted","Data":"bd8f3ac093e1410b5828b16d3a9d37f525658958b1fd3b2ccf0684450138faac"} Oct 09 10:46:11 crc kubenswrapper[4923]: I1009 10:46:11.600858 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:46:11 crc kubenswrapper[4923]: E1009 10:46:11.601166 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:46:25 crc kubenswrapper[4923]: I1009 10:46:25.601838 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:46:25 crc kubenswrapper[4923]: E1009 10:46:25.602571 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:46:32 crc kubenswrapper[4923]: I1009 10:46:32.891050 4923 generic.go:334] "Generic (PLEG): container finished" podID="d8452ab6-f295-464d-bf20-c4a0d7f6ce64" containerID="002d3a92aa9a8de6d170ca2c674335c8ad7bfc1ee2668bbd2fe45216aeb57f16" exitCode=0 Oct 09 10:46:32 crc kubenswrapper[4923]: I1009 10:46:32.891198 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" event={"ID":"d8452ab6-f295-464d-bf20-c4a0d7f6ce64","Type":"ContainerDied","Data":"002d3a92aa9a8de6d170ca2c674335c8ad7bfc1ee2668bbd2fe45216aeb57f16"} Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.404352 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.463610 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-inventory\") pod \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.464408 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ceph\") pod \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.464458 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xbft\" (UniqueName: \"kubernetes.io/projected/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-kube-api-access-4xbft\") pod \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.464546 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ssh-key\") pod \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\" (UID: \"d8452ab6-f295-464d-bf20-c4a0d7f6ce64\") " Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.471456 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ceph" (OuterVolumeSpecName: "ceph") pod "d8452ab6-f295-464d-bf20-c4a0d7f6ce64" (UID: "d8452ab6-f295-464d-bf20-c4a0d7f6ce64"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.472040 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-kube-api-access-4xbft" (OuterVolumeSpecName: "kube-api-access-4xbft") pod "d8452ab6-f295-464d-bf20-c4a0d7f6ce64" (UID: "d8452ab6-f295-464d-bf20-c4a0d7f6ce64"). InnerVolumeSpecName "kube-api-access-4xbft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.512027 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-inventory" (OuterVolumeSpecName: "inventory") pod "d8452ab6-f295-464d-bf20-c4a0d7f6ce64" (UID: "d8452ab6-f295-464d-bf20-c4a0d7f6ce64"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.522357 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d8452ab6-f295-464d-bf20-c4a0d7f6ce64" (UID: "d8452ab6-f295-464d-bf20-c4a0d7f6ce64"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.568631 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.568681 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.568695 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xbft\" (UniqueName: \"kubernetes.io/projected/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-kube-api-access-4xbft\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.568715 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d8452ab6-f295-464d-bf20-c4a0d7f6ce64-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.909777 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" event={"ID":"d8452ab6-f295-464d-bf20-c4a0d7f6ce64","Type":"ContainerDied","Data":"bd8f3ac093e1410b5828b16d3a9d37f525658958b1fd3b2ccf0684450138faac"} Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.910175 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd8f3ac093e1410b5828b16d3a9d37f525658958b1fd3b2ccf0684450138faac" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.910148 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.995328 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq"] Oct 09 10:46:34 crc kubenswrapper[4923]: E1009 10:46:34.995818 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8452ab6-f295-464d-bf20-c4a0d7f6ce64" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.995845 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8452ab6-f295-464d-bf20-c4a0d7f6ce64" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.996044 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8452ab6-f295-464d-bf20-c4a0d7f6ce64" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.996787 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.999201 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.999676 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.999733 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:46:34 crc kubenswrapper[4923]: I1009 10:46:34.999768 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.000263 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.014878 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq"] Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.079145 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.079230 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.079259 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.079291 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km6zs\" (UniqueName: \"kubernetes.io/projected/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-kube-api-access-km6zs\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.181322 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.181481 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.181528 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.181589 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km6zs\" (UniqueName: \"kubernetes.io/projected/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-kube-api-access-km6zs\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.187133 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.187236 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.192496 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.199383 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km6zs\" (UniqueName: \"kubernetes.io/projected/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-kube-api-access-km6zs\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-sncpq\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.316861 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.868920 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq"] Oct 09 10:46:35 crc kubenswrapper[4923]: I1009 10:46:35.922689 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" event={"ID":"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63","Type":"ContainerStarted","Data":"f2ae050f4a375e925ce5d77ba347fcc6455fc3191d477750bab0da8753c0c46f"} Oct 09 10:46:36 crc kubenswrapper[4923]: I1009 10:46:36.602258 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:46:36 crc kubenswrapper[4923]: E1009 10:46:36.602896 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:46:36 crc kubenswrapper[4923]: I1009 10:46:36.933908 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" event={"ID":"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63","Type":"ContainerStarted","Data":"5ca57616b99811837aa3267344f1d711dd69e4704d1e3a356ced881cc4ef7ac2"} Oct 09 10:46:36 crc kubenswrapper[4923]: I1009 10:46:36.966583 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" podStartSLOduration=2.547457092 podStartE2EDuration="2.966553484s" podCreationTimestamp="2025-10-09 10:46:34 +0000 UTC" firstStartedPulling="2025-10-09 10:46:35.876811352 +0000 UTC m=+2481.944993108" lastFinishedPulling="2025-10-09 10:46:36.295907754 +0000 UTC m=+2482.364089500" observedRunningTime="2025-10-09 10:46:36.953712298 +0000 UTC m=+2483.021894054" watchObservedRunningTime="2025-10-09 10:46:36.966553484 +0000 UTC m=+2483.034735250" Oct 09 10:46:41 crc kubenswrapper[4923]: I1009 10:46:41.983515 4923 generic.go:334] "Generic (PLEG): container finished" podID="b1ab7440-9c9d-4ba4-92c3-fe19bccbde63" containerID="5ca57616b99811837aa3267344f1d711dd69e4704d1e3a356ced881cc4ef7ac2" exitCode=0 Oct 09 10:46:41 crc kubenswrapper[4923]: I1009 10:46:41.983591 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" event={"ID":"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63","Type":"ContainerDied","Data":"5ca57616b99811837aa3267344f1d711dd69e4704d1e3a356ced881cc4ef7ac2"} Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.392090 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.468629 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ssh-key\") pod \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.468861 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-inventory\") pod \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.468893 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km6zs\" (UniqueName: \"kubernetes.io/projected/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-kube-api-access-km6zs\") pod \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.468985 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ceph\") pod \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\" (UID: \"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63\") " Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.476070 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-kube-api-access-km6zs" (OuterVolumeSpecName: "kube-api-access-km6zs") pod "b1ab7440-9c9d-4ba4-92c3-fe19bccbde63" (UID: "b1ab7440-9c9d-4ba4-92c3-fe19bccbde63"). InnerVolumeSpecName "kube-api-access-km6zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.477326 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ceph" (OuterVolumeSpecName: "ceph") pod "b1ab7440-9c9d-4ba4-92c3-fe19bccbde63" (UID: "b1ab7440-9c9d-4ba4-92c3-fe19bccbde63"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.500804 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-inventory" (OuterVolumeSpecName: "inventory") pod "b1ab7440-9c9d-4ba4-92c3-fe19bccbde63" (UID: "b1ab7440-9c9d-4ba4-92c3-fe19bccbde63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.501260 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b1ab7440-9c9d-4ba4-92c3-fe19bccbde63" (UID: "b1ab7440-9c9d-4ba4-92c3-fe19bccbde63"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.571675 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.571713 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.571724 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:43 crc kubenswrapper[4923]: I1009 10:46:43.571734 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km6zs\" (UniqueName: \"kubernetes.io/projected/b1ab7440-9c9d-4ba4-92c3-fe19bccbde63-kube-api-access-km6zs\") on node \"crc\" DevicePath \"\"" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.003817 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" event={"ID":"b1ab7440-9c9d-4ba4-92c3-fe19bccbde63","Type":"ContainerDied","Data":"f2ae050f4a375e925ce5d77ba347fcc6455fc3191d477750bab0da8753c0c46f"} Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.004339 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2ae050f4a375e925ce5d77ba347fcc6455fc3191d477750bab0da8753c0c46f" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.003946 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-sncpq" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.100848 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc"] Oct 09 10:46:44 crc kubenswrapper[4923]: E1009 10:46:44.101658 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1ab7440-9c9d-4ba4-92c3-fe19bccbde63" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.101779 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1ab7440-9c9d-4ba4-92c3-fe19bccbde63" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.102109 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1ab7440-9c9d-4ba4-92c3-fe19bccbde63" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.103035 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.107161 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.107211 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.107688 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.110550 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.114445 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.123427 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc"] Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.185152 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.185229 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.185284 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc8p8\" (UniqueName: \"kubernetes.io/projected/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-kube-api-access-lc8p8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.185351 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.287194 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.287303 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.287350 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.287400 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc8p8\" (UniqueName: \"kubernetes.io/projected/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-kube-api-access-lc8p8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.291882 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.292406 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.292973 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.307339 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc8p8\" (UniqueName: \"kubernetes.io/projected/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-kube-api-access-lc8p8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-fk6lc\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:44 crc kubenswrapper[4923]: I1009 10:46:44.424260 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:46:45 crc kubenswrapper[4923]: I1009 10:46:45.015648 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc"] Oct 09 10:46:46 crc kubenswrapper[4923]: I1009 10:46:46.029452 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" event={"ID":"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f","Type":"ContainerStarted","Data":"9f224ceb8352f26cb248010ec43c61fff7327dd1ea417305c22084f2c9c6a80e"} Oct 09 10:46:46 crc kubenswrapper[4923]: I1009 10:46:46.030170 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" event={"ID":"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f","Type":"ContainerStarted","Data":"a994ad556229b4e5c167e1be8d53f123bbc2c482cddf5428084e2218fc47038b"} Oct 09 10:46:46 crc kubenswrapper[4923]: I1009 10:46:46.059105 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" podStartSLOduration=1.487478703 podStartE2EDuration="2.05907823s" podCreationTimestamp="2025-10-09 10:46:44 +0000 UTC" firstStartedPulling="2025-10-09 10:46:45.027930401 +0000 UTC m=+2491.096112157" lastFinishedPulling="2025-10-09 10:46:45.599529928 +0000 UTC m=+2491.667711684" observedRunningTime="2025-10-09 10:46:46.054161163 +0000 UTC m=+2492.122342919" watchObservedRunningTime="2025-10-09 10:46:46.05907823 +0000 UTC m=+2492.127259986" Oct 09 10:46:50 crc kubenswrapper[4923]: I1009 10:46:50.601620 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:46:50 crc kubenswrapper[4923]: E1009 10:46:50.602662 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:47:03 crc kubenswrapper[4923]: I1009 10:47:03.602446 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:47:04 crc kubenswrapper[4923]: I1009 10:47:04.226022 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"1c6d2bbeaf151ee38b4f95dddf10b8f8656fa2db4c8e7968d5f9b745ca10fdcf"} Oct 09 10:47:22 crc kubenswrapper[4923]: I1009 10:47:22.395619 4923 generic.go:334] "Generic (PLEG): container finished" podID="ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f" containerID="9f224ceb8352f26cb248010ec43c61fff7327dd1ea417305c22084f2c9c6a80e" exitCode=0 Oct 09 10:47:22 crc kubenswrapper[4923]: I1009 10:47:22.395851 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" event={"ID":"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f","Type":"ContainerDied","Data":"9f224ceb8352f26cb248010ec43c61fff7327dd1ea417305c22084f2c9c6a80e"} Oct 09 10:47:23 crc kubenswrapper[4923]: I1009 10:47:23.867028 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:47:23 crc kubenswrapper[4923]: I1009 10:47:23.969339 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ssh-key\") pod \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " Oct 09 10:47:23 crc kubenswrapper[4923]: I1009 10:47:23.969527 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lc8p8\" (UniqueName: \"kubernetes.io/projected/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-kube-api-access-lc8p8\") pod \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " Oct 09 10:47:23 crc kubenswrapper[4923]: I1009 10:47:23.969679 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-inventory\") pod \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " Oct 09 10:47:23 crc kubenswrapper[4923]: I1009 10:47:23.969791 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ceph\") pod \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\" (UID: \"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f\") " Oct 09 10:47:23 crc kubenswrapper[4923]: I1009 10:47:23.979490 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ceph" (OuterVolumeSpecName: "ceph") pod "ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f" (UID: "ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:47:23 crc kubenswrapper[4923]: I1009 10:47:23.980829 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-kube-api-access-lc8p8" (OuterVolumeSpecName: "kube-api-access-lc8p8") pod "ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f" (UID: "ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f"). InnerVolumeSpecName "kube-api-access-lc8p8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.003805 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-inventory" (OuterVolumeSpecName: "inventory") pod "ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f" (UID: "ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.004734 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f" (UID: "ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.072463 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.072510 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lc8p8\" (UniqueName: \"kubernetes.io/projected/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-kube-api-access-lc8p8\") on node \"crc\" DevicePath \"\"" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.072560 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.072607 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.421826 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" event={"ID":"ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f","Type":"ContainerDied","Data":"a994ad556229b4e5c167e1be8d53f123bbc2c482cddf5428084e2218fc47038b"} Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.422462 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a994ad556229b4e5c167e1be8d53f123bbc2c482cddf5428084e2218fc47038b" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.421945 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-fk6lc" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.517409 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk"] Oct 09 10:47:24 crc kubenswrapper[4923]: E1009 10:47:24.518425 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.518542 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.518853 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.519955 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.525342 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.525406 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.525796 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.525996 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.526473 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.536720 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk"] Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.685531 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.686275 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.686509 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.686557 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs7pk\" (UniqueName: \"kubernetes.io/projected/18edca22-01b1-4d52-ac43-d93611eb5464-kube-api-access-rs7pk\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.789505 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.789646 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.789775 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.789814 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs7pk\" (UniqueName: \"kubernetes.io/projected/18edca22-01b1-4d52-ac43-d93611eb5464-kube-api-access-rs7pk\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.794720 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.797381 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.802833 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.811391 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs7pk\" (UniqueName: \"kubernetes.io/projected/18edca22-01b1-4d52-ac43-d93611eb5464-kube-api-access-rs7pk\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:24 crc kubenswrapper[4923]: I1009 10:47:24.847929 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:25 crc kubenswrapper[4923]: I1009 10:47:25.430792 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk"] Oct 09 10:47:26 crc kubenswrapper[4923]: I1009 10:47:26.446965 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" event={"ID":"18edca22-01b1-4d52-ac43-d93611eb5464","Type":"ContainerStarted","Data":"583dc6b728029808c6d274ed2cd1d8a8d10d2e99a6ce36a4c4cf1eddca85f7d1"} Oct 09 10:47:26 crc kubenswrapper[4923]: I1009 10:47:26.448920 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" event={"ID":"18edca22-01b1-4d52-ac43-d93611eb5464","Type":"ContainerStarted","Data":"cef080f59fe2152f949285fe60df7524387ca4d80d5fa125d74f691c65c50dab"} Oct 09 10:47:26 crc kubenswrapper[4923]: I1009 10:47:26.469237 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" podStartSLOduration=1.77163597 podStartE2EDuration="2.469188776s" podCreationTimestamp="2025-10-09 10:47:24 +0000 UTC" firstStartedPulling="2025-10-09 10:47:25.440987919 +0000 UTC m=+2531.509169675" lastFinishedPulling="2025-10-09 10:47:26.138540725 +0000 UTC m=+2532.206722481" observedRunningTime="2025-10-09 10:47:26.46429223 +0000 UTC m=+2532.532474006" watchObservedRunningTime="2025-10-09 10:47:26.469188776 +0000 UTC m=+2532.537370532" Oct 09 10:47:30 crc kubenswrapper[4923]: I1009 10:47:30.488967 4923 generic.go:334] "Generic (PLEG): container finished" podID="18edca22-01b1-4d52-ac43-d93611eb5464" containerID="583dc6b728029808c6d274ed2cd1d8a8d10d2e99a6ce36a4c4cf1eddca85f7d1" exitCode=0 Oct 09 10:47:30 crc kubenswrapper[4923]: I1009 10:47:30.489199 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" event={"ID":"18edca22-01b1-4d52-ac43-d93611eb5464","Type":"ContainerDied","Data":"583dc6b728029808c6d274ed2cd1d8a8d10d2e99a6ce36a4c4cf1eddca85f7d1"} Oct 09 10:47:31 crc kubenswrapper[4923]: I1009 10:47:31.946568 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:31 crc kubenswrapper[4923]: I1009 10:47:31.953246 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rs7pk\" (UniqueName: \"kubernetes.io/projected/18edca22-01b1-4d52-ac43-d93611eb5464-kube-api-access-rs7pk\") pod \"18edca22-01b1-4d52-ac43-d93611eb5464\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " Oct 09 10:47:31 crc kubenswrapper[4923]: I1009 10:47:31.953332 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ceph\") pod \"18edca22-01b1-4d52-ac43-d93611eb5464\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " Oct 09 10:47:31 crc kubenswrapper[4923]: I1009 10:47:31.953479 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ssh-key\") pod \"18edca22-01b1-4d52-ac43-d93611eb5464\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " Oct 09 10:47:31 crc kubenswrapper[4923]: I1009 10:47:31.953506 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-inventory\") pod \"18edca22-01b1-4d52-ac43-d93611eb5464\" (UID: \"18edca22-01b1-4d52-ac43-d93611eb5464\") " Oct 09 10:47:31 crc kubenswrapper[4923]: I1009 10:47:31.962732 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18edca22-01b1-4d52-ac43-d93611eb5464-kube-api-access-rs7pk" (OuterVolumeSpecName: "kube-api-access-rs7pk") pod "18edca22-01b1-4d52-ac43-d93611eb5464" (UID: "18edca22-01b1-4d52-ac43-d93611eb5464"). InnerVolumeSpecName "kube-api-access-rs7pk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:47:31 crc kubenswrapper[4923]: I1009 10:47:31.968183 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ceph" (OuterVolumeSpecName: "ceph") pod "18edca22-01b1-4d52-ac43-d93611eb5464" (UID: "18edca22-01b1-4d52-ac43-d93611eb5464"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:47:31 crc kubenswrapper[4923]: I1009 10:47:31.988701 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-inventory" (OuterVolumeSpecName: "inventory") pod "18edca22-01b1-4d52-ac43-d93611eb5464" (UID: "18edca22-01b1-4d52-ac43-d93611eb5464"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.004650 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18edca22-01b1-4d52-ac43-d93611eb5464" (UID: "18edca22-01b1-4d52-ac43-d93611eb5464"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.062503 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.062547 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.062561 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rs7pk\" (UniqueName: \"kubernetes.io/projected/18edca22-01b1-4d52-ac43-d93611eb5464-kube-api-access-rs7pk\") on node \"crc\" DevicePath \"\"" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.062580 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18edca22-01b1-4d52-ac43-d93611eb5464-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.510725 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" event={"ID":"18edca22-01b1-4d52-ac43-d93611eb5464","Type":"ContainerDied","Data":"cef080f59fe2152f949285fe60df7524387ca4d80d5fa125d74f691c65c50dab"} Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.510820 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cef080f59fe2152f949285fe60df7524387ca4d80d5fa125d74f691c65c50dab" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.510836 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.615459 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68"] Oct 09 10:47:32 crc kubenswrapper[4923]: E1009 10:47:32.615867 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18edca22-01b1-4d52-ac43-d93611eb5464" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.615893 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="18edca22-01b1-4d52-ac43-d93611eb5464" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.616128 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="18edca22-01b1-4d52-ac43-d93611eb5464" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.616919 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.620044 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.622610 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.622804 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.627069 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.627616 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.629734 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68"] Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.776533 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.776626 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmkzb\" (UniqueName: \"kubernetes.io/projected/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-kube-api-access-xmkzb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.776691 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.776789 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.878208 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.878671 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmkzb\" (UniqueName: \"kubernetes.io/projected/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-kube-api-access-xmkzb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.878729 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.878814 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.884784 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.885473 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.887716 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.898997 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmkzb\" (UniqueName: \"kubernetes.io/projected/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-kube-api-access-xmkzb\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-gjw68\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:32 crc kubenswrapper[4923]: I1009 10:47:32.943481 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:47:33 crc kubenswrapper[4923]: I1009 10:47:33.502549 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68"] Oct 09 10:47:33 crc kubenswrapper[4923]: I1009 10:47:33.522217 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" event={"ID":"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925","Type":"ContainerStarted","Data":"a32df8b8ce7504d4885b2d7c410ecc5ebc6ac5bea3285f28edc3b51cbeb2ac46"} Oct 09 10:47:34 crc kubenswrapper[4923]: I1009 10:47:34.535624 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" event={"ID":"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925","Type":"ContainerStarted","Data":"c02b5a23225e2a5e3a0de2cd383b73123216f52bb7bd29ff19faeca85b663868"} Oct 09 10:47:34 crc kubenswrapper[4923]: I1009 10:47:34.563003 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" podStartSLOduration=2.113707463 podStartE2EDuration="2.56297623s" podCreationTimestamp="2025-10-09 10:47:32 +0000 UTC" firstStartedPulling="2025-10-09 10:47:33.511237481 +0000 UTC m=+2539.579419237" lastFinishedPulling="2025-10-09 10:47:33.960506248 +0000 UTC m=+2540.028688004" observedRunningTime="2025-10-09 10:47:34.555020389 +0000 UTC m=+2540.623202145" watchObservedRunningTime="2025-10-09 10:47:34.56297623 +0000 UTC m=+2540.631157986" Oct 09 10:48:16 crc kubenswrapper[4923]: I1009 10:48:16.929518 4923 generic.go:334] "Generic (PLEG): container finished" podID="c66ce1f0-4c37-4e5d-ae58-bbebdfa66925" containerID="c02b5a23225e2a5e3a0de2cd383b73123216f52bb7bd29ff19faeca85b663868" exitCode=0 Oct 09 10:48:16 crc kubenswrapper[4923]: I1009 10:48:16.929576 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" event={"ID":"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925","Type":"ContainerDied","Data":"c02b5a23225e2a5e3a0de2cd383b73123216f52bb7bd29ff19faeca85b663868"} Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.412631 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.606160 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-inventory\") pod \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.606293 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmkzb\" (UniqueName: \"kubernetes.io/projected/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-kube-api-access-xmkzb\") pod \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.606355 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ceph\") pod \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.607616 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ssh-key\") pod \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\" (UID: \"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925\") " Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.615146 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ceph" (OuterVolumeSpecName: "ceph") pod "c66ce1f0-4c37-4e5d-ae58-bbebdfa66925" (UID: "c66ce1f0-4c37-4e5d-ae58-bbebdfa66925"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.617608 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-kube-api-access-xmkzb" (OuterVolumeSpecName: "kube-api-access-xmkzb") pod "c66ce1f0-4c37-4e5d-ae58-bbebdfa66925" (UID: "c66ce1f0-4c37-4e5d-ae58-bbebdfa66925"). InnerVolumeSpecName "kube-api-access-xmkzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.645177 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c66ce1f0-4c37-4e5d-ae58-bbebdfa66925" (UID: "c66ce1f0-4c37-4e5d-ae58-bbebdfa66925"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.653019 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-inventory" (OuterVolumeSpecName: "inventory") pod "c66ce1f0-4c37-4e5d-ae58-bbebdfa66925" (UID: "c66ce1f0-4c37-4e5d-ae58-bbebdfa66925"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.712084 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.712533 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.712558 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmkzb\" (UniqueName: \"kubernetes.io/projected/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-kube-api-access-xmkzb\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.712574 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c66ce1f0-4c37-4e5d-ae58-bbebdfa66925-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.949557 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" event={"ID":"c66ce1f0-4c37-4e5d-ae58-bbebdfa66925","Type":"ContainerDied","Data":"a32df8b8ce7504d4885b2d7c410ecc5ebc6ac5bea3285f28edc3b51cbeb2ac46"} Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.949627 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a32df8b8ce7504d4885b2d7c410ecc5ebc6ac5bea3285f28edc3b51cbeb2ac46" Oct 09 10:48:18 crc kubenswrapper[4923]: I1009 10:48:18.949766 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-gjw68" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.047064 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-c28nb"] Oct 09 10:48:19 crc kubenswrapper[4923]: E1009 10:48:19.047623 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c66ce1f0-4c37-4e5d-ae58-bbebdfa66925" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.047652 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c66ce1f0-4c37-4e5d-ae58-bbebdfa66925" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.048022 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c66ce1f0-4c37-4e5d-ae58-bbebdfa66925" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.050014 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.053865 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.053865 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.054958 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.054967 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.056551 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.059158 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-c28nb"] Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.228791 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.228905 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs2hn\" (UniqueName: \"kubernetes.io/projected/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-kube-api-access-hs2hn\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.228954 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ceph\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.229011 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.330699 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs2hn\" (UniqueName: \"kubernetes.io/projected/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-kube-api-access-hs2hn\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.330790 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ceph\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.330850 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.330956 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.337670 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.337923 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ceph\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.339498 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.363493 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs2hn\" (UniqueName: \"kubernetes.io/projected/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-kube-api-access-hs2hn\") pod \"ssh-known-hosts-edpm-deployment-c28nb\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.368925 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.920581 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-c28nb"] Oct 09 10:48:19 crc kubenswrapper[4923]: I1009 10:48:19.961325 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" event={"ID":"7fc44ca7-2ff1-4677-b3d1-860332fef1c4","Type":"ContainerStarted","Data":"4c9db2c23973837defef0a2e7190da928cec1d86862ca558e789259039e97bec"} Oct 09 10:48:20 crc kubenswrapper[4923]: I1009 10:48:20.973511 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" event={"ID":"7fc44ca7-2ff1-4677-b3d1-860332fef1c4","Type":"ContainerStarted","Data":"d381d312cff9ee487cd76ce9ef5a710e1d3ef6456ce4cb661fb05e1e04525372"} Oct 09 10:48:20 crc kubenswrapper[4923]: I1009 10:48:20.996435 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" podStartSLOduration=1.5633638319999998 podStartE2EDuration="1.9963982s" podCreationTimestamp="2025-10-09 10:48:19 +0000 UTC" firstStartedPulling="2025-10-09 10:48:19.934786628 +0000 UTC m=+2586.002968384" lastFinishedPulling="2025-10-09 10:48:20.367821006 +0000 UTC m=+2586.436002752" observedRunningTime="2025-10-09 10:48:20.993417938 +0000 UTC m=+2587.061599694" watchObservedRunningTime="2025-10-09 10:48:20.9963982 +0000 UTC m=+2587.064579956" Oct 09 10:48:30 crc kubenswrapper[4923]: I1009 10:48:30.059118 4923 generic.go:334] "Generic (PLEG): container finished" podID="7fc44ca7-2ff1-4677-b3d1-860332fef1c4" containerID="d381d312cff9ee487cd76ce9ef5a710e1d3ef6456ce4cb661fb05e1e04525372" exitCode=0 Oct 09 10:48:30 crc kubenswrapper[4923]: I1009 10:48:30.059186 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" event={"ID":"7fc44ca7-2ff1-4677-b3d1-860332fef1c4","Type":"ContainerDied","Data":"d381d312cff9ee487cd76ce9ef5a710e1d3ef6456ce4cb661fb05e1e04525372"} Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.495426 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.601954 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ssh-key-openstack-edpm-ipam\") pod \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.602147 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ceph\") pod \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.602181 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs2hn\" (UniqueName: \"kubernetes.io/projected/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-kube-api-access-hs2hn\") pod \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.602232 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-inventory-0\") pod \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\" (UID: \"7fc44ca7-2ff1-4677-b3d1-860332fef1c4\") " Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.611062 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-kube-api-access-hs2hn" (OuterVolumeSpecName: "kube-api-access-hs2hn") pod "7fc44ca7-2ff1-4677-b3d1-860332fef1c4" (UID: "7fc44ca7-2ff1-4677-b3d1-860332fef1c4"). InnerVolumeSpecName "kube-api-access-hs2hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.611214 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ceph" (OuterVolumeSpecName: "ceph") pod "7fc44ca7-2ff1-4677-b3d1-860332fef1c4" (UID: "7fc44ca7-2ff1-4677-b3d1-860332fef1c4"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.629680 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7fc44ca7-2ff1-4677-b3d1-860332fef1c4" (UID: "7fc44ca7-2ff1-4677-b3d1-860332fef1c4"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.629850 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "7fc44ca7-2ff1-4677-b3d1-860332fef1c4" (UID: "7fc44ca7-2ff1-4677-b3d1-860332fef1c4"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.705432 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs2hn\" (UniqueName: \"kubernetes.io/projected/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-kube-api-access-hs2hn\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.705904 4923 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.705926 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:31 crc kubenswrapper[4923]: I1009 10:48:31.705945 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7fc44ca7-2ff1-4677-b3d1-860332fef1c4-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.082860 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" event={"ID":"7fc44ca7-2ff1-4677-b3d1-860332fef1c4","Type":"ContainerDied","Data":"4c9db2c23973837defef0a2e7190da928cec1d86862ca558e789259039e97bec"} Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.082918 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c9db2c23973837defef0a2e7190da928cec1d86862ca558e789259039e97bec" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.083356 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-c28nb" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.172697 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc"] Oct 09 10:48:32 crc kubenswrapper[4923]: E1009 10:48:32.173216 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc44ca7-2ff1-4677-b3d1-860332fef1c4" containerName="ssh-known-hosts-edpm-deployment" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.173238 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc44ca7-2ff1-4677-b3d1-860332fef1c4" containerName="ssh-known-hosts-edpm-deployment" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.173527 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc44ca7-2ff1-4677-b3d1-860332fef1c4" containerName="ssh-known-hosts-edpm-deployment" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.174532 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.180562 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.180905 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.181171 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.181513 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.181692 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.185424 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc"] Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.318910 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.318972 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.319108 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.319135 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2txmx\" (UniqueName: \"kubernetes.io/projected/6881fbdc-1adc-4922-aca4-484a5fa2a348-kube-api-access-2txmx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.420984 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.421035 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2txmx\" (UniqueName: \"kubernetes.io/projected/6881fbdc-1adc-4922-aca4-484a5fa2a348-kube-api-access-2txmx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.421126 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.421149 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.425798 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.426450 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.426717 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.444673 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2txmx\" (UniqueName: \"kubernetes.io/projected/6881fbdc-1adc-4922-aca4-484a5fa2a348-kube-api-access-2txmx\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-wjkcc\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:32 crc kubenswrapper[4923]: I1009 10:48:32.496327 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:33 crc kubenswrapper[4923]: I1009 10:48:33.072528 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc"] Oct 09 10:48:33 crc kubenswrapper[4923]: I1009 10:48:33.103012 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" event={"ID":"6881fbdc-1adc-4922-aca4-484a5fa2a348","Type":"ContainerStarted","Data":"b95634cef5f969fb0fe5feecdf6ae23407db0568b0d241b314447046e558ab78"} Oct 09 10:48:34 crc kubenswrapper[4923]: I1009 10:48:34.113179 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" event={"ID":"6881fbdc-1adc-4922-aca4-484a5fa2a348","Type":"ContainerStarted","Data":"cfd53cfee97714176d4f7dbe410c4b27ec12e3da740e44d1d16951ad450e946f"} Oct 09 10:48:34 crc kubenswrapper[4923]: I1009 10:48:34.142026 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" podStartSLOduration=1.710012371 podStartE2EDuration="2.141970739s" podCreationTimestamp="2025-10-09 10:48:32 +0000 UTC" firstStartedPulling="2025-10-09 10:48:33.091277979 +0000 UTC m=+2599.159459745" lastFinishedPulling="2025-10-09 10:48:33.523236327 +0000 UTC m=+2599.591418113" observedRunningTime="2025-10-09 10:48:34.131305413 +0000 UTC m=+2600.199487199" watchObservedRunningTime="2025-10-09 10:48:34.141970739 +0000 UTC m=+2600.210152495" Oct 09 10:48:41 crc kubenswrapper[4923]: I1009 10:48:41.184309 4923 generic.go:334] "Generic (PLEG): container finished" podID="6881fbdc-1adc-4922-aca4-484a5fa2a348" containerID="cfd53cfee97714176d4f7dbe410c4b27ec12e3da740e44d1d16951ad450e946f" exitCode=0 Oct 09 10:48:41 crc kubenswrapper[4923]: I1009 10:48:41.184371 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" event={"ID":"6881fbdc-1adc-4922-aca4-484a5fa2a348","Type":"ContainerDied","Data":"cfd53cfee97714176d4f7dbe410c4b27ec12e3da740e44d1d16951ad450e946f"} Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.671368 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.791051 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-inventory\") pod \"6881fbdc-1adc-4922-aca4-484a5fa2a348\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.791692 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ceph\") pod \"6881fbdc-1adc-4922-aca4-484a5fa2a348\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.791919 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ssh-key\") pod \"6881fbdc-1adc-4922-aca4-484a5fa2a348\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.791977 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2txmx\" (UniqueName: \"kubernetes.io/projected/6881fbdc-1adc-4922-aca4-484a5fa2a348-kube-api-access-2txmx\") pod \"6881fbdc-1adc-4922-aca4-484a5fa2a348\" (UID: \"6881fbdc-1adc-4922-aca4-484a5fa2a348\") " Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.799225 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ceph" (OuterVolumeSpecName: "ceph") pod "6881fbdc-1adc-4922-aca4-484a5fa2a348" (UID: "6881fbdc-1adc-4922-aca4-484a5fa2a348"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.800177 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6881fbdc-1adc-4922-aca4-484a5fa2a348-kube-api-access-2txmx" (OuterVolumeSpecName: "kube-api-access-2txmx") pod "6881fbdc-1adc-4922-aca4-484a5fa2a348" (UID: "6881fbdc-1adc-4922-aca4-484a5fa2a348"). InnerVolumeSpecName "kube-api-access-2txmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.822597 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6881fbdc-1adc-4922-aca4-484a5fa2a348" (UID: "6881fbdc-1adc-4922-aca4-484a5fa2a348"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.822988 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-inventory" (OuterVolumeSpecName: "inventory") pod "6881fbdc-1adc-4922-aca4-484a5fa2a348" (UID: "6881fbdc-1adc-4922-aca4-484a5fa2a348"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.894202 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.894241 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.894254 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6881fbdc-1adc-4922-aca4-484a5fa2a348-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:42 crc kubenswrapper[4923]: I1009 10:48:42.894265 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2txmx\" (UniqueName: \"kubernetes.io/projected/6881fbdc-1adc-4922-aca4-484a5fa2a348-kube-api-access-2txmx\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.206605 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" event={"ID":"6881fbdc-1adc-4922-aca4-484a5fa2a348","Type":"ContainerDied","Data":"b95634cef5f969fb0fe5feecdf6ae23407db0568b0d241b314447046e558ab78"} Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.206662 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b95634cef5f969fb0fe5feecdf6ae23407db0568b0d241b314447046e558ab78" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.207219 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-wjkcc" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.293742 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6"] Oct 09 10:48:43 crc kubenswrapper[4923]: E1009 10:48:43.294275 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6881fbdc-1adc-4922-aca4-484a5fa2a348" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.294310 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="6881fbdc-1adc-4922-aca4-484a5fa2a348" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.294533 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="6881fbdc-1adc-4922-aca4-484a5fa2a348" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.295437 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.297384 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.297409 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.297799 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.299735 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.300676 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhr24\" (UniqueName: \"kubernetes.io/projected/6494d91a-6ead-46aa-b724-20fbca0b5319-kube-api-access-jhr24\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.300734 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.300794 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.300820 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.305318 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.308195 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6"] Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.403110 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhr24\" (UniqueName: \"kubernetes.io/projected/6494d91a-6ead-46aa-b724-20fbca0b5319-kube-api-access-jhr24\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.403186 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.403234 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.403259 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.408452 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.408774 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.410207 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.424969 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhr24\" (UniqueName: \"kubernetes.io/projected/6494d91a-6ead-46aa-b724-20fbca0b5319-kube-api-access-jhr24\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:43 crc kubenswrapper[4923]: I1009 10:48:43.636882 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:44 crc kubenswrapper[4923]: I1009 10:48:44.202091 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6"] Oct 09 10:48:44 crc kubenswrapper[4923]: I1009 10:48:44.217107 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" event={"ID":"6494d91a-6ead-46aa-b724-20fbca0b5319","Type":"ContainerStarted","Data":"7f21a0a0614204adc3223a7f56a3866fad9d8db6f2504ef965326187d20d88f7"} Oct 09 10:48:45 crc kubenswrapper[4923]: I1009 10:48:45.228524 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" event={"ID":"6494d91a-6ead-46aa-b724-20fbca0b5319","Type":"ContainerStarted","Data":"10d12ce5bb2ee1b2ab52eb247ec1258913db7621b988162d96784c98f047e9b8"} Oct 09 10:48:45 crc kubenswrapper[4923]: I1009 10:48:45.256857 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" podStartSLOduration=1.7217462879999998 podStartE2EDuration="2.256824862s" podCreationTimestamp="2025-10-09 10:48:43 +0000 UTC" firstStartedPulling="2025-10-09 10:48:44.208272702 +0000 UTC m=+2610.276454458" lastFinishedPulling="2025-10-09 10:48:44.743351276 +0000 UTC m=+2610.811533032" observedRunningTime="2025-10-09 10:48:45.247798622 +0000 UTC m=+2611.315980398" watchObservedRunningTime="2025-10-09 10:48:45.256824862 +0000 UTC m=+2611.325006618" Oct 09 10:48:55 crc kubenswrapper[4923]: I1009 10:48:55.325828 4923 generic.go:334] "Generic (PLEG): container finished" podID="6494d91a-6ead-46aa-b724-20fbca0b5319" containerID="10d12ce5bb2ee1b2ab52eb247ec1258913db7621b988162d96784c98f047e9b8" exitCode=0 Oct 09 10:48:55 crc kubenswrapper[4923]: I1009 10:48:55.325946 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" event={"ID":"6494d91a-6ead-46aa-b724-20fbca0b5319","Type":"ContainerDied","Data":"10d12ce5bb2ee1b2ab52eb247ec1258913db7621b988162d96784c98f047e9b8"} Oct 09 10:48:56 crc kubenswrapper[4923]: I1009 10:48:56.783665 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:56 crc kubenswrapper[4923]: I1009 10:48:56.914870 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-inventory\") pod \"6494d91a-6ead-46aa-b724-20fbca0b5319\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " Oct 09 10:48:56 crc kubenswrapper[4923]: I1009 10:48:56.915070 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ssh-key\") pod \"6494d91a-6ead-46aa-b724-20fbca0b5319\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " Oct 09 10:48:56 crc kubenswrapper[4923]: I1009 10:48:56.915110 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhr24\" (UniqueName: \"kubernetes.io/projected/6494d91a-6ead-46aa-b724-20fbca0b5319-kube-api-access-jhr24\") pod \"6494d91a-6ead-46aa-b724-20fbca0b5319\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " Oct 09 10:48:56 crc kubenswrapper[4923]: I1009 10:48:56.915943 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ceph\") pod \"6494d91a-6ead-46aa-b724-20fbca0b5319\" (UID: \"6494d91a-6ead-46aa-b724-20fbca0b5319\") " Oct 09 10:48:56 crc kubenswrapper[4923]: I1009 10:48:56.923721 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6494d91a-6ead-46aa-b724-20fbca0b5319-kube-api-access-jhr24" (OuterVolumeSpecName: "kube-api-access-jhr24") pod "6494d91a-6ead-46aa-b724-20fbca0b5319" (UID: "6494d91a-6ead-46aa-b724-20fbca0b5319"). InnerVolumeSpecName "kube-api-access-jhr24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:48:56 crc kubenswrapper[4923]: I1009 10:48:56.924183 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ceph" (OuterVolumeSpecName: "ceph") pod "6494d91a-6ead-46aa-b724-20fbca0b5319" (UID: "6494d91a-6ead-46aa-b724-20fbca0b5319"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:56 crc kubenswrapper[4923]: I1009 10:48:56.945242 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-inventory" (OuterVolumeSpecName: "inventory") pod "6494d91a-6ead-46aa-b724-20fbca0b5319" (UID: "6494d91a-6ead-46aa-b724-20fbca0b5319"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:56 crc kubenswrapper[4923]: I1009 10:48:56.948936 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6494d91a-6ead-46aa-b724-20fbca0b5319" (UID: "6494d91a-6ead-46aa-b724-20fbca0b5319"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.020096 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.020162 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.020176 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6494d91a-6ead-46aa-b724-20fbca0b5319-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.020194 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhr24\" (UniqueName: \"kubernetes.io/projected/6494d91a-6ead-46aa-b724-20fbca0b5319-kube-api-access-jhr24\") on node \"crc\" DevicePath \"\"" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.350919 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" event={"ID":"6494d91a-6ead-46aa-b724-20fbca0b5319","Type":"ContainerDied","Data":"7f21a0a0614204adc3223a7f56a3866fad9d8db6f2504ef965326187d20d88f7"} Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.350992 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f21a0a0614204adc3223a7f56a3866fad9d8db6f2504ef965326187d20d88f7" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.351084 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.441499 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5"] Oct 09 10:48:57 crc kubenswrapper[4923]: E1009 10:48:57.442051 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6494d91a-6ead-46aa-b724-20fbca0b5319" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.442076 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="6494d91a-6ead-46aa-b724-20fbca0b5319" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.442283 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="6494d91a-6ead-46aa-b724-20fbca0b5319" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.443173 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.447580 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.447931 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.448063 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.448188 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.448384 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.448519 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.448661 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.451004 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.467303 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5"] Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.632596 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.632662 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.632691 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.632762 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvfjz\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-kube-api-access-mvfjz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.632792 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.632815 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.632840 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.632869 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.633209 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.633311 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.633381 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.633422 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.633676 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.735963 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.736661 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.736700 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.736734 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.736841 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvfjz\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-kube-api-access-mvfjz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.736905 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.736937 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.737148 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.737308 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.737378 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.737448 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.737499 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.737527 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.743279 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.743453 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.744269 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.744274 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.745261 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.746342 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.747688 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.747808 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.749144 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.750248 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.752878 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.753527 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.757097 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvfjz\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-kube-api-access-mvfjz\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:57 crc kubenswrapper[4923]: I1009 10:48:57.774854 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:48:58 crc kubenswrapper[4923]: I1009 10:48:58.332452 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5"] Oct 09 10:48:58 crc kubenswrapper[4923]: I1009 10:48:58.370691 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" event={"ID":"b8e09b1b-4163-4570-96c6-aca084f2ea50","Type":"ContainerStarted","Data":"8649840c61554e77db79e17d91567986aefc92f0321e53832c9c25e6cd098942"} Oct 09 10:48:59 crc kubenswrapper[4923]: I1009 10:48:59.383664 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" event={"ID":"b8e09b1b-4163-4570-96c6-aca084f2ea50","Type":"ContainerStarted","Data":"729309e60684d610abd162ef01ac813aef317288b0568ee9996f35e5fa6d9e9e"} Oct 09 10:48:59 crc kubenswrapper[4923]: I1009 10:48:59.412777 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" podStartSLOduration=1.854673641 podStartE2EDuration="2.412739104s" podCreationTimestamp="2025-10-09 10:48:57 +0000 UTC" firstStartedPulling="2025-10-09 10:48:58.359268965 +0000 UTC m=+2624.427450721" lastFinishedPulling="2025-10-09 10:48:58.917334428 +0000 UTC m=+2624.985516184" observedRunningTime="2025-10-09 10:48:59.405925744 +0000 UTC m=+2625.474107510" watchObservedRunningTime="2025-10-09 10:48:59.412739104 +0000 UTC m=+2625.480920850" Oct 09 10:49:24 crc kubenswrapper[4923]: I1009 10:49:24.600156 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:49:24 crc kubenswrapper[4923]: I1009 10:49:24.601099 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:49:30 crc kubenswrapper[4923]: I1009 10:49:30.678660 4923 generic.go:334] "Generic (PLEG): container finished" podID="b8e09b1b-4163-4570-96c6-aca084f2ea50" containerID="729309e60684d610abd162ef01ac813aef317288b0568ee9996f35e5fa6d9e9e" exitCode=0 Oct 09 10:49:30 crc kubenswrapper[4923]: I1009 10:49:30.678783 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" event={"ID":"b8e09b1b-4163-4570-96c6-aca084f2ea50","Type":"ContainerDied","Data":"729309e60684d610abd162ef01ac813aef317288b0568ee9996f35e5fa6d9e9e"} Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.162240 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.190950 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ovn-combined-ca-bundle\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191008 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-bootstrap-combined-ca-bundle\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191043 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ceph\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191164 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-neutron-metadata-combined-ca-bundle\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191229 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-ovn-default-certs-0\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191274 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191297 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-repo-setup-combined-ca-bundle\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191325 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvfjz\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-kube-api-access-mvfjz\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191374 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ssh-key\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191391 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-libvirt-combined-ca-bundle\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191428 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191471 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-inventory\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.191505 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-nova-combined-ca-bundle\") pod \"b8e09b1b-4163-4570-96c6-aca084f2ea50\" (UID: \"b8e09b1b-4163-4570-96c6-aca084f2ea50\") " Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.201358 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-kube-api-access-mvfjz" (OuterVolumeSpecName: "kube-api-access-mvfjz") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "kube-api-access-mvfjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.202094 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.203475 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.203508 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.204545 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.204888 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.205279 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ceph" (OuterVolumeSpecName: "ceph") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.206015 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.207271 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.208503 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.216201 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.234275 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.242414 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-inventory" (OuterVolumeSpecName: "inventory") pod "b8e09b1b-4163-4570-96c6-aca084f2ea50" (UID: "b8e09b1b-4163-4570-96c6-aca084f2ea50"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293456 4923 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293508 4923 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293527 4923 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293542 4923 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293579 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvfjz\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-kube-api-access-mvfjz\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293593 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293607 4923 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293622 4923 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/b8e09b1b-4163-4570-96c6-aca084f2ea50-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293636 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293648 4923 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293662 4923 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293674 4923 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.293685 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b8e09b1b-4163-4570-96c6-aca084f2ea50-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.701465 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" event={"ID":"b8e09b1b-4163-4570-96c6-aca084f2ea50","Type":"ContainerDied","Data":"8649840c61554e77db79e17d91567986aefc92f0321e53832c9c25e6cd098942"} Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.701517 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8649840c61554e77db79e17d91567986aefc92f0321e53832c9c25e6cd098942" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.701715 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.826011 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7"] Oct 09 10:49:32 crc kubenswrapper[4923]: E1009 10:49:32.826608 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e09b1b-4163-4570-96c6-aca084f2ea50" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.826633 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e09b1b-4163-4570-96c6-aca084f2ea50" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.826896 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e09b1b-4163-4570-96c6-aca084f2ea50" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.827670 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.830358 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.831378 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.831557 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.837936 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.845529 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.847524 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7"] Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.919455 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.919627 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.919661 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kszh\" (UniqueName: \"kubernetes.io/projected/43c42b69-73b7-4758-92cc-431d2837fb39-kube-api-access-2kszh\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:32 crc kubenswrapper[4923]: I1009 10:49:32.919697 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.021772 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.021896 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.021942 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kszh\" (UniqueName: \"kubernetes.io/projected/43c42b69-73b7-4758-92cc-431d2837fb39-kube-api-access-2kszh\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.021988 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.032645 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.033323 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.037648 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.045766 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kszh\" (UniqueName: \"kubernetes.io/projected/43c42b69-73b7-4758-92cc-431d2837fb39-kube-api-access-2kszh\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.150620 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:33 crc kubenswrapper[4923]: I1009 10:49:33.727520 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7"] Oct 09 10:49:34 crc kubenswrapper[4923]: I1009 10:49:34.722744 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" event={"ID":"43c42b69-73b7-4758-92cc-431d2837fb39","Type":"ContainerStarted","Data":"ee47a4c1779dfc0948689076b57fb241a3f5198e31caefc1b2ec0fc1e279fcac"} Oct 09 10:49:34 crc kubenswrapper[4923]: I1009 10:49:34.723329 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" event={"ID":"43c42b69-73b7-4758-92cc-431d2837fb39","Type":"ContainerStarted","Data":"f669ae9564ed474a35b3f6126611cb09940e34ce1a78a1b2872deb1d22c46b61"} Oct 09 10:49:34 crc kubenswrapper[4923]: I1009 10:49:34.751416 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" podStartSLOduration=2.29895765 podStartE2EDuration="2.751391795s" podCreationTimestamp="2025-10-09 10:49:32 +0000 UTC" firstStartedPulling="2025-10-09 10:49:33.738892363 +0000 UTC m=+2659.807074119" lastFinishedPulling="2025-10-09 10:49:34.191326508 +0000 UTC m=+2660.259508264" observedRunningTime="2025-10-09 10:49:34.739840795 +0000 UTC m=+2660.808022551" watchObservedRunningTime="2025-10-09 10:49:34.751391795 +0000 UTC m=+2660.819573551" Oct 09 10:49:40 crc kubenswrapper[4923]: I1009 10:49:40.791067 4923 generic.go:334] "Generic (PLEG): container finished" podID="43c42b69-73b7-4758-92cc-431d2837fb39" containerID="ee47a4c1779dfc0948689076b57fb241a3f5198e31caefc1b2ec0fc1e279fcac" exitCode=0 Oct 09 10:49:40 crc kubenswrapper[4923]: I1009 10:49:40.791150 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" event={"ID":"43c42b69-73b7-4758-92cc-431d2837fb39","Type":"ContainerDied","Data":"ee47a4c1779dfc0948689076b57fb241a3f5198e31caefc1b2ec0fc1e279fcac"} Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.234202 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.259551 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-inventory\") pod \"43c42b69-73b7-4758-92cc-431d2837fb39\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.259621 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ssh-key\") pod \"43c42b69-73b7-4758-92cc-431d2837fb39\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.259734 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ceph\") pod \"43c42b69-73b7-4758-92cc-431d2837fb39\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.259946 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kszh\" (UniqueName: \"kubernetes.io/projected/43c42b69-73b7-4758-92cc-431d2837fb39-kube-api-access-2kszh\") pod \"43c42b69-73b7-4758-92cc-431d2837fb39\" (UID: \"43c42b69-73b7-4758-92cc-431d2837fb39\") " Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.281831 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ceph" (OuterVolumeSpecName: "ceph") pod "43c42b69-73b7-4758-92cc-431d2837fb39" (UID: "43c42b69-73b7-4758-92cc-431d2837fb39"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.282808 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c42b69-73b7-4758-92cc-431d2837fb39-kube-api-access-2kszh" (OuterVolumeSpecName: "kube-api-access-2kszh") pod "43c42b69-73b7-4758-92cc-431d2837fb39" (UID: "43c42b69-73b7-4758-92cc-431d2837fb39"). InnerVolumeSpecName "kube-api-access-2kszh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.298523 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "43c42b69-73b7-4758-92cc-431d2837fb39" (UID: "43c42b69-73b7-4758-92cc-431d2837fb39"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.299447 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-inventory" (OuterVolumeSpecName: "inventory") pod "43c42b69-73b7-4758-92cc-431d2837fb39" (UID: "43c42b69-73b7-4758-92cc-431d2837fb39"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.363844 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kszh\" (UniqueName: \"kubernetes.io/projected/43c42b69-73b7-4758-92cc-431d2837fb39-kube-api-access-2kszh\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.364313 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.364328 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.364341 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/43c42b69-73b7-4758-92cc-431d2837fb39-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.813867 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" event={"ID":"43c42b69-73b7-4758-92cc-431d2837fb39","Type":"ContainerDied","Data":"f669ae9564ed474a35b3f6126611cb09940e34ce1a78a1b2872deb1d22c46b61"} Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.813927 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f669ae9564ed474a35b3f6126611cb09940e34ce1a78a1b2872deb1d22c46b61" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.814003 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.898641 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq"] Oct 09 10:49:42 crc kubenswrapper[4923]: E1009 10:49:42.899251 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43c42b69-73b7-4758-92cc-431d2837fb39" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.899275 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="43c42b69-73b7-4758-92cc-431d2837fb39" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.899509 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="43c42b69-73b7-4758-92cc-431d2837fb39" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.900464 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.904319 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.904570 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.908160 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.908169 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.918065 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.919388 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.925109 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq"] Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.977821 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.977887 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/eedf18b1-d421-493d-838c-965b961d21a3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.978180 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttv6c\" (UniqueName: \"kubernetes.io/projected/eedf18b1-d421-493d-838c-965b961d21a3-kube-api-access-ttv6c\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.978467 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.978651 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:42 crc kubenswrapper[4923]: I1009 10:49:42.978706 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.081037 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttv6c\" (UniqueName: \"kubernetes.io/projected/eedf18b1-d421-493d-838c-965b961d21a3-kube-api-access-ttv6c\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.081167 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.081234 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.081266 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.081357 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.081389 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/eedf18b1-d421-493d-838c-965b961d21a3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.082454 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/eedf18b1-d421-493d-838c-965b961d21a3-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.088881 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.089106 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.092578 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.094340 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.125689 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttv6c\" (UniqueName: \"kubernetes.io/projected/eedf18b1-d421-493d-838c-965b961d21a3-kube-api-access-ttv6c\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-z8dvq\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.220252 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.808522 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq"] Oct 09 10:49:43 crc kubenswrapper[4923]: I1009 10:49:43.827385 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" event={"ID":"eedf18b1-d421-493d-838c-965b961d21a3","Type":"ContainerStarted","Data":"d03330b135f05d1d1ae36ee5b2205d10c3d696a3b93525d2d2010269711ba04d"} Oct 09 10:49:44 crc kubenswrapper[4923]: I1009 10:49:44.839625 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" event={"ID":"eedf18b1-d421-493d-838c-965b961d21a3","Type":"ContainerStarted","Data":"9ce51bb7222b6f0b64fa2aadf5da4e98629a28149ccc4445a45cf03951484c8b"} Oct 09 10:49:44 crc kubenswrapper[4923]: I1009 10:49:44.878544 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" podStartSLOduration=2.450197378 podStartE2EDuration="2.878509874s" podCreationTimestamp="2025-10-09 10:49:42 +0000 UTC" firstStartedPulling="2025-10-09 10:49:43.817986283 +0000 UTC m=+2669.886168039" lastFinishedPulling="2025-10-09 10:49:44.246298779 +0000 UTC m=+2670.314480535" observedRunningTime="2025-10-09 10:49:44.863105868 +0000 UTC m=+2670.931287634" watchObservedRunningTime="2025-10-09 10:49:44.878509874 +0000 UTC m=+2670.946691630" Oct 09 10:49:54 crc kubenswrapper[4923]: I1009 10:49:54.599568 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:49:54 crc kubenswrapper[4923]: I1009 10:49:54.600378 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:50:24 crc kubenswrapper[4923]: I1009 10:50:24.599401 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:50:24 crc kubenswrapper[4923]: I1009 10:50:24.600154 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:50:24 crc kubenswrapper[4923]: I1009 10:50:24.600225 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:50:24 crc kubenswrapper[4923]: I1009 10:50:24.601217 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1c6d2bbeaf151ee38b4f95dddf10b8f8656fa2db4c8e7968d5f9b745ca10fdcf"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:50:24 crc kubenswrapper[4923]: I1009 10:50:24.601287 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://1c6d2bbeaf151ee38b4f95dddf10b8f8656fa2db4c8e7968d5f9b745ca10fdcf" gracePeriod=600 Oct 09 10:50:25 crc kubenswrapper[4923]: I1009 10:50:25.262561 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="1c6d2bbeaf151ee38b4f95dddf10b8f8656fa2db4c8e7968d5f9b745ca10fdcf" exitCode=0 Oct 09 10:50:25 crc kubenswrapper[4923]: I1009 10:50:25.262638 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"1c6d2bbeaf151ee38b4f95dddf10b8f8656fa2db4c8e7968d5f9b745ca10fdcf"} Oct 09 10:50:25 crc kubenswrapper[4923]: I1009 10:50:25.263462 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735"} Oct 09 10:50:25 crc kubenswrapper[4923]: I1009 10:50:25.263498 4923 scope.go:117] "RemoveContainer" containerID="ba758ebfcd2188729947b05ad4569cd50c515196b6318bdaed3f9e63c59b8391" Oct 09 10:50:53 crc kubenswrapper[4923]: I1009 10:50:53.533110 4923 generic.go:334] "Generic (PLEG): container finished" podID="eedf18b1-d421-493d-838c-965b961d21a3" containerID="9ce51bb7222b6f0b64fa2aadf5da4e98629a28149ccc4445a45cf03951484c8b" exitCode=0 Oct 09 10:50:53 crc kubenswrapper[4923]: I1009 10:50:53.533218 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" event={"ID":"eedf18b1-d421-493d-838c-965b961d21a3","Type":"ContainerDied","Data":"9ce51bb7222b6f0b64fa2aadf5da4e98629a28149ccc4445a45cf03951484c8b"} Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.047507 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.243498 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/eedf18b1-d421-493d-838c-965b961d21a3-ovncontroller-config-0\") pod \"eedf18b1-d421-493d-838c-965b961d21a3\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.244021 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttv6c\" (UniqueName: \"kubernetes.io/projected/eedf18b1-d421-493d-838c-965b961d21a3-kube-api-access-ttv6c\") pod \"eedf18b1-d421-493d-838c-965b961d21a3\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.244091 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ceph\") pod \"eedf18b1-d421-493d-838c-965b961d21a3\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.244170 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ssh-key\") pod \"eedf18b1-d421-493d-838c-965b961d21a3\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.244293 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ovn-combined-ca-bundle\") pod \"eedf18b1-d421-493d-838c-965b961d21a3\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.244365 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-inventory\") pod \"eedf18b1-d421-493d-838c-965b961d21a3\" (UID: \"eedf18b1-d421-493d-838c-965b961d21a3\") " Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.251230 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ceph" (OuterVolumeSpecName: "ceph") pod "eedf18b1-d421-493d-838c-965b961d21a3" (UID: "eedf18b1-d421-493d-838c-965b961d21a3"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.265121 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eedf18b1-d421-493d-838c-965b961d21a3-kube-api-access-ttv6c" (OuterVolumeSpecName: "kube-api-access-ttv6c") pod "eedf18b1-d421-493d-838c-965b961d21a3" (UID: "eedf18b1-d421-493d-838c-965b961d21a3"). InnerVolumeSpecName "kube-api-access-ttv6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.269830 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "eedf18b1-d421-493d-838c-965b961d21a3" (UID: "eedf18b1-d421-493d-838c-965b961d21a3"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.290993 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eedf18b1-d421-493d-838c-965b961d21a3" (UID: "eedf18b1-d421-493d-838c-965b961d21a3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.292457 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedf18b1-d421-493d-838c-965b961d21a3-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "eedf18b1-d421-493d-838c-965b961d21a3" (UID: "eedf18b1-d421-493d-838c-965b961d21a3"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.295315 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-inventory" (OuterVolumeSpecName: "inventory") pod "eedf18b1-d421-493d-838c-965b961d21a3" (UID: "eedf18b1-d421-493d-838c-965b961d21a3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.349511 4923 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/eedf18b1-d421-493d-838c-965b961d21a3-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.349557 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttv6c\" (UniqueName: \"kubernetes.io/projected/eedf18b1-d421-493d-838c-965b961d21a3-kube-api-access-ttv6c\") on node \"crc\" DevicePath \"\"" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.349574 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.349587 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.349600 4923 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.349612 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eedf18b1-d421-493d-838c-965b961d21a3-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.555376 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" event={"ID":"eedf18b1-d421-493d-838c-965b961d21a3","Type":"ContainerDied","Data":"d03330b135f05d1d1ae36ee5b2205d10c3d696a3b93525d2d2010269711ba04d"} Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.555465 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d03330b135f05d1d1ae36ee5b2205d10c3d696a3b93525d2d2010269711ba04d" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.555418 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-z8dvq" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.674463 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp"] Oct 09 10:50:55 crc kubenswrapper[4923]: E1009 10:50:55.675004 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedf18b1-d421-493d-838c-965b961d21a3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.675025 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedf18b1-d421-493d-838c-965b961d21a3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.675525 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="eedf18b1-d421-493d-838c-965b961d21a3" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.676352 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.679883 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.679925 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.680148 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.680247 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.680336 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.680566 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.683250 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.696570 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp"] Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.861536 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.861629 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.862619 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.862705 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.862742 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.862900 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4thb\" (UniqueName: \"kubernetes.io/projected/480a61a2-d67f-45ed-b2ef-d24b8b750035-kube-api-access-j4thb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.863147 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.966379 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.966506 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.966543 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.966603 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.966639 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.966672 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.966736 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4thb\" (UniqueName: \"kubernetes.io/projected/480a61a2-d67f-45ed-b2ef-d24b8b750035-kube-api-access-j4thb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.971675 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.971719 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.972381 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.972814 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.973973 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.974521 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:55 crc kubenswrapper[4923]: I1009 10:50:55.991268 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4thb\" (UniqueName: \"kubernetes.io/projected/480a61a2-d67f-45ed-b2ef-d24b8b750035-kube-api-access-j4thb\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:56 crc kubenswrapper[4923]: I1009 10:50:56.003147 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:50:56 crc kubenswrapper[4923]: I1009 10:50:56.679441 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp"] Oct 09 10:50:57 crc kubenswrapper[4923]: I1009 10:50:57.579493 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" event={"ID":"480a61a2-d67f-45ed-b2ef-d24b8b750035","Type":"ContainerStarted","Data":"95aa657cdadb42afaeb2beb91c796ff1cbc3328f7a08607c188858b19876c303"} Oct 09 10:50:58 crc kubenswrapper[4923]: I1009 10:50:58.591939 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" event={"ID":"480a61a2-d67f-45ed-b2ef-d24b8b750035","Type":"ContainerStarted","Data":"7a34dcfe88f511224c9d5032cf842c004d94b16f779c5180bd0b4e7606b45a77"} Oct 09 10:50:58 crc kubenswrapper[4923]: I1009 10:50:58.625659 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" podStartSLOduration=2.572227861 podStartE2EDuration="3.625633436s" podCreationTimestamp="2025-10-09 10:50:55 +0000 UTC" firstStartedPulling="2025-10-09 10:50:56.693102435 +0000 UTC m=+2742.761284191" lastFinishedPulling="2025-10-09 10:50:57.74650801 +0000 UTC m=+2743.814689766" observedRunningTime="2025-10-09 10:50:58.621129342 +0000 UTC m=+2744.689311118" watchObservedRunningTime="2025-10-09 10:50:58.625633436 +0000 UTC m=+2744.693815192" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.242484 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b4psb"] Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.245508 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.258690 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b4psb"] Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.348289 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfl7r\" (UniqueName: \"kubernetes.io/projected/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-kube-api-access-zfl7r\") pod \"community-operators-b4psb\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.348368 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-utilities\") pod \"community-operators-b4psb\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.348436 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-catalog-content\") pod \"community-operators-b4psb\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.450294 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfl7r\" (UniqueName: \"kubernetes.io/projected/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-kube-api-access-zfl7r\") pod \"community-operators-b4psb\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.450391 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-utilities\") pod \"community-operators-b4psb\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.450457 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-catalog-content\") pod \"community-operators-b4psb\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.451112 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-utilities\") pod \"community-operators-b4psb\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.451148 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-catalog-content\") pod \"community-operators-b4psb\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.478150 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfl7r\" (UniqueName: \"kubernetes.io/projected/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-kube-api-access-zfl7r\") pod \"community-operators-b4psb\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:50:59 crc kubenswrapper[4923]: I1009 10:50:59.573486 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:51:00 crc kubenswrapper[4923]: I1009 10:51:00.060621 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b4psb"] Oct 09 10:51:00 crc kubenswrapper[4923]: I1009 10:51:00.614348 4923 generic.go:334] "Generic (PLEG): container finished" podID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerID="c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf" exitCode=0 Oct 09 10:51:00 crc kubenswrapper[4923]: I1009 10:51:00.619025 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4psb" event={"ID":"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19","Type":"ContainerDied","Data":"c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf"} Oct 09 10:51:00 crc kubenswrapper[4923]: I1009 10:51:00.619084 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4psb" event={"ID":"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19","Type":"ContainerStarted","Data":"467aa17f4f72cbd5b284a80333e03af76ba7c52d16840522eedd1a7eb8ccba9d"} Oct 09 10:51:02 crc kubenswrapper[4923]: I1009 10:51:02.636184 4923 generic.go:334] "Generic (PLEG): container finished" podID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerID="a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1" exitCode=0 Oct 09 10:51:02 crc kubenswrapper[4923]: I1009 10:51:02.636335 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4psb" event={"ID":"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19","Type":"ContainerDied","Data":"a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1"} Oct 09 10:51:03 crc kubenswrapper[4923]: I1009 10:51:03.651721 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4psb" event={"ID":"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19","Type":"ContainerStarted","Data":"019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3"} Oct 09 10:51:03 crc kubenswrapper[4923]: I1009 10:51:03.680598 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b4psb" podStartSLOduration=2.218519125 podStartE2EDuration="4.680565097s" podCreationTimestamp="2025-10-09 10:50:59 +0000 UTC" firstStartedPulling="2025-10-09 10:51:00.616616028 +0000 UTC m=+2746.684797784" lastFinishedPulling="2025-10-09 10:51:03.078662 +0000 UTC m=+2749.146843756" observedRunningTime="2025-10-09 10:51:03.671574268 +0000 UTC m=+2749.739756024" watchObservedRunningTime="2025-10-09 10:51:03.680565097 +0000 UTC m=+2749.748746853" Oct 09 10:51:09 crc kubenswrapper[4923]: I1009 10:51:09.575307 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:51:09 crc kubenswrapper[4923]: I1009 10:51:09.576137 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:51:09 crc kubenswrapper[4923]: I1009 10:51:09.629343 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:51:09 crc kubenswrapper[4923]: I1009 10:51:09.761274 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:51:09 crc kubenswrapper[4923]: I1009 10:51:09.872482 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b4psb"] Oct 09 10:51:11 crc kubenswrapper[4923]: I1009 10:51:11.730648 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b4psb" podUID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerName="registry-server" containerID="cri-o://019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3" gracePeriod=2 Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.219722 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.278811 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7kb87"] Oct 09 10:51:12 crc kubenswrapper[4923]: E1009 10:51:12.279686 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerName="extract-content" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.279818 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerName="extract-content" Oct 09 10:51:12 crc kubenswrapper[4923]: E1009 10:51:12.279910 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerName="registry-server" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.279975 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerName="registry-server" Oct 09 10:51:12 crc kubenswrapper[4923]: E1009 10:51:12.280067 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerName="extract-utilities" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.280157 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerName="extract-utilities" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.280459 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerName="registry-server" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.288275 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.300549 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kb87"] Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.399178 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-catalog-content\") pod \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.405055 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-utilities\") pod \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.405132 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfl7r\" (UniqueName: \"kubernetes.io/projected/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-kube-api-access-zfl7r\") pod \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\" (UID: \"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19\") " Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.405434 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-utilities\") pod \"redhat-marketplace-7kb87\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.405476 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-catalog-content\") pod \"redhat-marketplace-7kb87\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.405594 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crcs7\" (UniqueName: \"kubernetes.io/projected/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-kube-api-access-crcs7\") pod \"redhat-marketplace-7kb87\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.406588 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-utilities" (OuterVolumeSpecName: "utilities") pod "e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" (UID: "e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.415026 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-kube-api-access-zfl7r" (OuterVolumeSpecName: "kube-api-access-zfl7r") pod "e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" (UID: "e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19"). InnerVolumeSpecName "kube-api-access-zfl7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.495239 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" (UID: "e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.507844 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crcs7\" (UniqueName: \"kubernetes.io/projected/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-kube-api-access-crcs7\") pod \"redhat-marketplace-7kb87\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.508119 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-utilities\") pod \"redhat-marketplace-7kb87\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.508232 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-catalog-content\") pod \"redhat-marketplace-7kb87\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.508555 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.508577 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.508589 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfl7r\" (UniqueName: \"kubernetes.io/projected/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19-kube-api-access-zfl7r\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.509158 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-catalog-content\") pod \"redhat-marketplace-7kb87\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.509960 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-utilities\") pod \"redhat-marketplace-7kb87\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.531827 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crcs7\" (UniqueName: \"kubernetes.io/projected/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-kube-api-access-crcs7\") pod \"redhat-marketplace-7kb87\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.617903 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.759501 4923 generic.go:334] "Generic (PLEG): container finished" podID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" containerID="019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3" exitCode=0 Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.759568 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4psb" event={"ID":"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19","Type":"ContainerDied","Data":"019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3"} Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.759606 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b4psb" event={"ID":"e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19","Type":"ContainerDied","Data":"467aa17f4f72cbd5b284a80333e03af76ba7c52d16840522eedd1a7eb8ccba9d"} Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.759628 4923 scope.go:117] "RemoveContainer" containerID="019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.759833 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b4psb" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.792958 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b4psb"] Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.802701 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b4psb"] Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.802937 4923 scope.go:117] "RemoveContainer" containerID="a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.829320 4923 scope.go:117] "RemoveContainer" containerID="c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.863726 4923 scope.go:117] "RemoveContainer" containerID="019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3" Oct 09 10:51:12 crc kubenswrapper[4923]: E1009 10:51:12.864320 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3\": container with ID starting with 019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3 not found: ID does not exist" containerID="019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.864364 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3"} err="failed to get container status \"019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3\": rpc error: code = NotFound desc = could not find container \"019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3\": container with ID starting with 019ad043344ef5eb93998b7701ca4d0fcb592fa1c2188021ad795b6c197c90a3 not found: ID does not exist" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.864392 4923 scope.go:117] "RemoveContainer" containerID="a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1" Oct 09 10:51:12 crc kubenswrapper[4923]: E1009 10:51:12.864677 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1\": container with ID starting with a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1 not found: ID does not exist" containerID="a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.864713 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1"} err="failed to get container status \"a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1\": rpc error: code = NotFound desc = could not find container \"a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1\": container with ID starting with a05398a5490d9e1133923eb9466eb34c44d71befd7a9a485067c028c5db8f1b1 not found: ID does not exist" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.864728 4923 scope.go:117] "RemoveContainer" containerID="c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf" Oct 09 10:51:12 crc kubenswrapper[4923]: E1009 10:51:12.865030 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf\": container with ID starting with c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf not found: ID does not exist" containerID="c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf" Oct 09 10:51:12 crc kubenswrapper[4923]: I1009 10:51:12.865057 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf"} err="failed to get container status \"c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf\": rpc error: code = NotFound desc = could not find container \"c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf\": container with ID starting with c4c56ad0953d8cf2f76ed39f54ab3e000343bb9d06a8f67d5e04f7d3bfe52cbf not found: ID does not exist" Oct 09 10:51:13 crc kubenswrapper[4923]: I1009 10:51:13.112777 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kb87"] Oct 09 10:51:13 crc kubenswrapper[4923]: I1009 10:51:13.771529 4923 generic.go:334] "Generic (PLEG): container finished" podID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerID="1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305" exitCode=0 Oct 09 10:51:13 crc kubenswrapper[4923]: I1009 10:51:13.771612 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kb87" event={"ID":"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d","Type":"ContainerDied","Data":"1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305"} Oct 09 10:51:13 crc kubenswrapper[4923]: I1009 10:51:13.772204 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kb87" event={"ID":"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d","Type":"ContainerStarted","Data":"7fbae57ac70a0e004f36eb190ddbaeb2329ecaf760a11d086376f3c041fe9817"} Oct 09 10:51:13 crc kubenswrapper[4923]: I1009 10:51:13.776985 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 10:51:14 crc kubenswrapper[4923]: I1009 10:51:14.617203 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19" path="/var/lib/kubelet/pods/e59cc9c4-4ddb-4d2f-b7f7-fbce55a0af19/volumes" Oct 09 10:51:14 crc kubenswrapper[4923]: I1009 10:51:14.786865 4923 generic.go:334] "Generic (PLEG): container finished" podID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerID="56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108" exitCode=0 Oct 09 10:51:14 crc kubenswrapper[4923]: I1009 10:51:14.786924 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kb87" event={"ID":"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d","Type":"ContainerDied","Data":"56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108"} Oct 09 10:51:15 crc kubenswrapper[4923]: I1009 10:51:15.800034 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kb87" event={"ID":"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d","Type":"ContainerStarted","Data":"35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725"} Oct 09 10:51:15 crc kubenswrapper[4923]: I1009 10:51:15.825430 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7kb87" podStartSLOduration=2.30379768 podStartE2EDuration="3.825400057s" podCreationTimestamp="2025-10-09 10:51:12 +0000 UTC" firstStartedPulling="2025-10-09 10:51:13.776630925 +0000 UTC m=+2759.844812691" lastFinishedPulling="2025-10-09 10:51:15.298233302 +0000 UTC m=+2761.366415068" observedRunningTime="2025-10-09 10:51:15.820426379 +0000 UTC m=+2761.888608135" watchObservedRunningTime="2025-10-09 10:51:15.825400057 +0000 UTC m=+2761.893581813" Oct 09 10:51:22 crc kubenswrapper[4923]: I1009 10:51:22.619388 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:22 crc kubenswrapper[4923]: I1009 10:51:22.620378 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:22 crc kubenswrapper[4923]: I1009 10:51:22.673794 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:22 crc kubenswrapper[4923]: I1009 10:51:22.919373 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:22 crc kubenswrapper[4923]: I1009 10:51:22.984112 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kb87"] Oct 09 10:51:24 crc kubenswrapper[4923]: I1009 10:51:24.884244 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7kb87" podUID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerName="registry-server" containerID="cri-o://35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725" gracePeriod=2 Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.373174 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.486193 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-utilities\") pod \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.486337 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crcs7\" (UniqueName: \"kubernetes.io/projected/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-kube-api-access-crcs7\") pod \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.486361 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-catalog-content\") pod \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\" (UID: \"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d\") " Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.487325 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-utilities" (OuterVolumeSpecName: "utilities") pod "8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" (UID: "8e7b78ad-53b6-4312-84cc-b68f6ca1c47d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.493127 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-kube-api-access-crcs7" (OuterVolumeSpecName: "kube-api-access-crcs7") pod "8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" (UID: "8e7b78ad-53b6-4312-84cc-b68f6ca1c47d"). InnerVolumeSpecName "kube-api-access-crcs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.505781 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" (UID: "8e7b78ad-53b6-4312-84cc-b68f6ca1c47d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.589187 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.589611 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crcs7\" (UniqueName: \"kubernetes.io/projected/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-kube-api-access-crcs7\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.589699 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.899090 4923 generic.go:334] "Generic (PLEG): container finished" podID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerID="35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725" exitCode=0 Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.899449 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kb87" event={"ID":"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d","Type":"ContainerDied","Data":"35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725"} Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.899481 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7kb87" event={"ID":"8e7b78ad-53b6-4312-84cc-b68f6ca1c47d","Type":"ContainerDied","Data":"7fbae57ac70a0e004f36eb190ddbaeb2329ecaf760a11d086376f3c041fe9817"} Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.899499 4923 scope.go:117] "RemoveContainer" containerID="35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.899631 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7kb87" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.940719 4923 scope.go:117] "RemoveContainer" containerID="56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108" Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.951345 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kb87"] Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.962647 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7kb87"] Oct 09 10:51:25 crc kubenswrapper[4923]: I1009 10:51:25.967199 4923 scope.go:117] "RemoveContainer" containerID="1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305" Oct 09 10:51:26 crc kubenswrapper[4923]: I1009 10:51:26.016271 4923 scope.go:117] "RemoveContainer" containerID="35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725" Oct 09 10:51:26 crc kubenswrapper[4923]: E1009 10:51:26.018526 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725\": container with ID starting with 35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725 not found: ID does not exist" containerID="35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725" Oct 09 10:51:26 crc kubenswrapper[4923]: I1009 10:51:26.018594 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725"} err="failed to get container status \"35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725\": rpc error: code = NotFound desc = could not find container \"35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725\": container with ID starting with 35a9e2a57110a25bb437e7caf5b9a4941e5de2736d20336042c0dce9bb922725 not found: ID does not exist" Oct 09 10:51:26 crc kubenswrapper[4923]: I1009 10:51:26.018644 4923 scope.go:117] "RemoveContainer" containerID="56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108" Oct 09 10:51:26 crc kubenswrapper[4923]: E1009 10:51:26.019332 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108\": container with ID starting with 56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108 not found: ID does not exist" containerID="56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108" Oct 09 10:51:26 crc kubenswrapper[4923]: I1009 10:51:26.019423 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108"} err="failed to get container status \"56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108\": rpc error: code = NotFound desc = could not find container \"56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108\": container with ID starting with 56d59c244b35d8256194bd5f7b73d36b6d39b700c641ff5a3dfde979186e1108 not found: ID does not exist" Oct 09 10:51:26 crc kubenswrapper[4923]: I1009 10:51:26.019468 4923 scope.go:117] "RemoveContainer" containerID="1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305" Oct 09 10:51:26 crc kubenswrapper[4923]: E1009 10:51:26.020035 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305\": container with ID starting with 1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305 not found: ID does not exist" containerID="1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305" Oct 09 10:51:26 crc kubenswrapper[4923]: I1009 10:51:26.020082 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305"} err="failed to get container status \"1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305\": rpc error: code = NotFound desc = could not find container \"1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305\": container with ID starting with 1282d5de0e50c501dbb4e348e62f860c1ca01c8c0f5f882dad584b6c0c18c305 not found: ID does not exist" Oct 09 10:51:26 crc kubenswrapper[4923]: I1009 10:51:26.614048 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" path="/var/lib/kubelet/pods/8e7b78ad-53b6-4312-84cc-b68f6ca1c47d/volumes" Oct 09 10:51:56 crc kubenswrapper[4923]: I1009 10:51:56.191487 4923 generic.go:334] "Generic (PLEG): container finished" podID="480a61a2-d67f-45ed-b2ef-d24b8b750035" containerID="7a34dcfe88f511224c9d5032cf842c004d94b16f779c5180bd0b4e7606b45a77" exitCode=0 Oct 09 10:51:56 crc kubenswrapper[4923]: I1009 10:51:56.191539 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" event={"ID":"480a61a2-d67f-45ed-b2ef-d24b8b750035","Type":"ContainerDied","Data":"7a34dcfe88f511224c9d5032cf842c004d94b16f779c5180bd0b4e7606b45a77"} Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.629667 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.756178 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ceph\") pod \"480a61a2-d67f-45ed-b2ef-d24b8b750035\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.756311 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-ovn-metadata-agent-neutron-config-0\") pod \"480a61a2-d67f-45ed-b2ef-d24b8b750035\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.756377 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-inventory\") pod \"480a61a2-d67f-45ed-b2ef-d24b8b750035\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.756530 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-metadata-combined-ca-bundle\") pod \"480a61a2-d67f-45ed-b2ef-d24b8b750035\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.756625 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ssh-key\") pod \"480a61a2-d67f-45ed-b2ef-d24b8b750035\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.757577 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4thb\" (UniqueName: \"kubernetes.io/projected/480a61a2-d67f-45ed-b2ef-d24b8b750035-kube-api-access-j4thb\") pod \"480a61a2-d67f-45ed-b2ef-d24b8b750035\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.757789 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-nova-metadata-neutron-config-0\") pod \"480a61a2-d67f-45ed-b2ef-d24b8b750035\" (UID: \"480a61a2-d67f-45ed-b2ef-d24b8b750035\") " Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.763377 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/480a61a2-d67f-45ed-b2ef-d24b8b750035-kube-api-access-j4thb" (OuterVolumeSpecName: "kube-api-access-j4thb") pod "480a61a2-d67f-45ed-b2ef-d24b8b750035" (UID: "480a61a2-d67f-45ed-b2ef-d24b8b750035"). InnerVolumeSpecName "kube-api-access-j4thb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.763951 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "480a61a2-d67f-45ed-b2ef-d24b8b750035" (UID: "480a61a2-d67f-45ed-b2ef-d24b8b750035"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.764079 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ceph" (OuterVolumeSpecName: "ceph") pod "480a61a2-d67f-45ed-b2ef-d24b8b750035" (UID: "480a61a2-d67f-45ed-b2ef-d24b8b750035"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.790635 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "480a61a2-d67f-45ed-b2ef-d24b8b750035" (UID: "480a61a2-d67f-45ed-b2ef-d24b8b750035"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.790733 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "480a61a2-d67f-45ed-b2ef-d24b8b750035" (UID: "480a61a2-d67f-45ed-b2ef-d24b8b750035"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.791398 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "480a61a2-d67f-45ed-b2ef-d24b8b750035" (UID: "480a61a2-d67f-45ed-b2ef-d24b8b750035"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.791523 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-inventory" (OuterVolumeSpecName: "inventory") pod "480a61a2-d67f-45ed-b2ef-d24b8b750035" (UID: "480a61a2-d67f-45ed-b2ef-d24b8b750035"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.861161 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.861203 4923 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.861220 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.861233 4923 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.861247 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.861260 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4thb\" (UniqueName: \"kubernetes.io/projected/480a61a2-d67f-45ed-b2ef-d24b8b750035-kube-api-access-j4thb\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:57 crc kubenswrapper[4923]: I1009 10:51:57.861276 4923 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/480a61a2-d67f-45ed-b2ef-d24b8b750035-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.218143 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" event={"ID":"480a61a2-d67f-45ed-b2ef-d24b8b750035","Type":"ContainerDied","Data":"95aa657cdadb42afaeb2beb91c796ff1cbc3328f7a08607c188858b19876c303"} Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.218684 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95aa657cdadb42afaeb2beb91c796ff1cbc3328f7a08607c188858b19876c303" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.218226 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.338093 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f"] Oct 09 10:51:58 crc kubenswrapper[4923]: E1009 10:51:58.338690 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerName="extract-utilities" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.338709 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerName="extract-utilities" Oct 09 10:51:58 crc kubenswrapper[4923]: E1009 10:51:58.338736 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerName="registry-server" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.338744 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerName="registry-server" Oct 09 10:51:58 crc kubenswrapper[4923]: E1009 10:51:58.338788 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="480a61a2-d67f-45ed-b2ef-d24b8b750035" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.338798 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="480a61a2-d67f-45ed-b2ef-d24b8b750035" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 09 10:51:58 crc kubenswrapper[4923]: E1009 10:51:58.338822 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerName="extract-content" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.338830 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerName="extract-content" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.339076 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="480a61a2-d67f-45ed-b2ef-d24b8b750035" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.339097 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e7b78ad-53b6-4312-84cc-b68f6ca1c47d" containerName="registry-server" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.340028 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.344490 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.345287 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.345513 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.345730 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.346059 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.346205 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.349526 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f"] Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.490441 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.490504 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.490557 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.490694 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7m9f2\" (UniqueName: \"kubernetes.io/projected/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-kube-api-access-7m9f2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.490844 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.490919 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.592646 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.592879 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.592917 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.593004 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.593043 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7m9f2\" (UniqueName: \"kubernetes.io/projected/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-kube-api-access-7m9f2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.593296 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.598983 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.599768 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.600317 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.601916 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.603268 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.612943 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7m9f2\" (UniqueName: \"kubernetes.io/projected/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-kube-api-access-7m9f2\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2f26f\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:58 crc kubenswrapper[4923]: I1009 10:51:58.671411 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:51:59 crc kubenswrapper[4923]: I1009 10:51:59.234035 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f"] Oct 09 10:52:00 crc kubenswrapper[4923]: I1009 10:52:00.243382 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" event={"ID":"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45","Type":"ContainerStarted","Data":"cc6c8052c31e87f89c711e596e77b24658d30852d2a5ad081605cbe6853fc576"} Oct 09 10:52:00 crc kubenswrapper[4923]: I1009 10:52:00.243725 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" event={"ID":"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45","Type":"ContainerStarted","Data":"047550eec0873de661d50ab9faef876712212d632c921450f35971751b5dfeff"} Oct 09 10:52:00 crc kubenswrapper[4923]: I1009 10:52:00.262605 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" podStartSLOduration=1.75910473 podStartE2EDuration="2.26258069s" podCreationTimestamp="2025-10-09 10:51:58 +0000 UTC" firstStartedPulling="2025-10-09 10:51:59.246246151 +0000 UTC m=+2805.314427917" lastFinishedPulling="2025-10-09 10:51:59.749722131 +0000 UTC m=+2805.817903877" observedRunningTime="2025-10-09 10:52:00.260329257 +0000 UTC m=+2806.328511023" watchObservedRunningTime="2025-10-09 10:52:00.26258069 +0000 UTC m=+2806.330762446" Oct 09 10:52:24 crc kubenswrapper[4923]: I1009 10:52:24.599594 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:52:24 crc kubenswrapper[4923]: I1009 10:52:24.600432 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:52:54 crc kubenswrapper[4923]: I1009 10:52:54.599800 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:52:54 crc kubenswrapper[4923]: I1009 10:52:54.600462 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:53:24 crc kubenswrapper[4923]: I1009 10:53:24.599832 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 10:53:24 crc kubenswrapper[4923]: I1009 10:53:24.600706 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 10:53:24 crc kubenswrapper[4923]: I1009 10:53:24.615437 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 10:53:24 crc kubenswrapper[4923]: I1009 10:53:24.616412 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 10:53:24 crc kubenswrapper[4923]: I1009 10:53:24.616551 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" gracePeriod=600 Oct 09 10:53:24 crc kubenswrapper[4923]: E1009 10:53:24.743323 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.123003 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" exitCode=0 Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.123198 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735"} Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.123435 4923 scope.go:117] "RemoveContainer" containerID="1c6d2bbeaf151ee38b4f95dddf10b8f8656fa2db4c8e7968d5f9b745ca10fdcf" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.124328 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:53:25 crc kubenswrapper[4923]: E1009 10:53:25.124623 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.491558 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pf7d9"] Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.494354 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.524195 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pf7d9"] Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.583154 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-catalog-content\") pod \"redhat-operators-pf7d9\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.583244 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlhzs\" (UniqueName: \"kubernetes.io/projected/d5782118-cbff-484d-a3fc-6be69adb37bb-kube-api-access-zlhzs\") pod \"redhat-operators-pf7d9\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.583417 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-utilities\") pod \"redhat-operators-pf7d9\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.686198 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-utilities\") pod \"redhat-operators-pf7d9\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.686789 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-catalog-content\") pod \"redhat-operators-pf7d9\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.686855 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-utilities\") pod \"redhat-operators-pf7d9\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.686973 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlhzs\" (UniqueName: \"kubernetes.io/projected/d5782118-cbff-484d-a3fc-6be69adb37bb-kube-api-access-zlhzs\") pod \"redhat-operators-pf7d9\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.687163 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-catalog-content\") pod \"redhat-operators-pf7d9\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.736151 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlhzs\" (UniqueName: \"kubernetes.io/projected/d5782118-cbff-484d-a3fc-6be69adb37bb-kube-api-access-zlhzs\") pod \"redhat-operators-pf7d9\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:25 crc kubenswrapper[4923]: I1009 10:53:25.826597 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:26 crc kubenswrapper[4923]: I1009 10:53:26.326246 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pf7d9"] Oct 09 10:53:27 crc kubenswrapper[4923]: I1009 10:53:27.155112 4923 generic.go:334] "Generic (PLEG): container finished" podID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerID="bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02" exitCode=0 Oct 09 10:53:27 crc kubenswrapper[4923]: I1009 10:53:27.155349 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pf7d9" event={"ID":"d5782118-cbff-484d-a3fc-6be69adb37bb","Type":"ContainerDied","Data":"bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02"} Oct 09 10:53:27 crc kubenswrapper[4923]: I1009 10:53:27.155482 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pf7d9" event={"ID":"d5782118-cbff-484d-a3fc-6be69adb37bb","Type":"ContainerStarted","Data":"4958a7c0e176dd14e3f8a3479ea76ca29862c81e6d839f25d5ee7ed458029c4f"} Oct 09 10:53:28 crc kubenswrapper[4923]: I1009 10:53:28.168976 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pf7d9" event={"ID":"d5782118-cbff-484d-a3fc-6be69adb37bb","Type":"ContainerStarted","Data":"7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e"} Oct 09 10:53:29 crc kubenswrapper[4923]: I1009 10:53:29.182899 4923 generic.go:334] "Generic (PLEG): container finished" podID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerID="7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e" exitCode=0 Oct 09 10:53:29 crc kubenswrapper[4923]: I1009 10:53:29.182917 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pf7d9" event={"ID":"d5782118-cbff-484d-a3fc-6be69adb37bb","Type":"ContainerDied","Data":"7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e"} Oct 09 10:53:30 crc kubenswrapper[4923]: I1009 10:53:30.197724 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pf7d9" event={"ID":"d5782118-cbff-484d-a3fc-6be69adb37bb","Type":"ContainerStarted","Data":"2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53"} Oct 09 10:53:30 crc kubenswrapper[4923]: I1009 10:53:30.227057 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pf7d9" podStartSLOduration=2.739430724 podStartE2EDuration="5.227033674s" podCreationTimestamp="2025-10-09 10:53:25 +0000 UTC" firstStartedPulling="2025-10-09 10:53:27.158428497 +0000 UTC m=+2893.226610253" lastFinishedPulling="2025-10-09 10:53:29.646031447 +0000 UTC m=+2895.714213203" observedRunningTime="2025-10-09 10:53:30.219020172 +0000 UTC m=+2896.287201948" watchObservedRunningTime="2025-10-09 10:53:30.227033674 +0000 UTC m=+2896.295215430" Oct 09 10:53:35 crc kubenswrapper[4923]: I1009 10:53:35.827234 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:35 crc kubenswrapper[4923]: I1009 10:53:35.828047 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:35 crc kubenswrapper[4923]: I1009 10:53:35.875294 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:36 crc kubenswrapper[4923]: I1009 10:53:36.311598 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:36 crc kubenswrapper[4923]: I1009 10:53:36.374859 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pf7d9"] Oct 09 10:53:38 crc kubenswrapper[4923]: I1009 10:53:38.278801 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pf7d9" podUID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerName="registry-server" containerID="cri-o://2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53" gracePeriod=2 Oct 09 10:53:38 crc kubenswrapper[4923]: I1009 10:53:38.809935 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:38 crc kubenswrapper[4923]: I1009 10:53:38.910411 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zlhzs\" (UniqueName: \"kubernetes.io/projected/d5782118-cbff-484d-a3fc-6be69adb37bb-kube-api-access-zlhzs\") pod \"d5782118-cbff-484d-a3fc-6be69adb37bb\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " Oct 09 10:53:38 crc kubenswrapper[4923]: I1009 10:53:38.910679 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-catalog-content\") pod \"d5782118-cbff-484d-a3fc-6be69adb37bb\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " Oct 09 10:53:38 crc kubenswrapper[4923]: I1009 10:53:38.910786 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-utilities\") pod \"d5782118-cbff-484d-a3fc-6be69adb37bb\" (UID: \"d5782118-cbff-484d-a3fc-6be69adb37bb\") " Oct 09 10:53:38 crc kubenswrapper[4923]: I1009 10:53:38.912046 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-utilities" (OuterVolumeSpecName: "utilities") pod "d5782118-cbff-484d-a3fc-6be69adb37bb" (UID: "d5782118-cbff-484d-a3fc-6be69adb37bb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:53:38 crc kubenswrapper[4923]: I1009 10:53:38.919436 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5782118-cbff-484d-a3fc-6be69adb37bb-kube-api-access-zlhzs" (OuterVolumeSpecName: "kube-api-access-zlhzs") pod "d5782118-cbff-484d-a3fc-6be69adb37bb" (UID: "d5782118-cbff-484d-a3fc-6be69adb37bb"). InnerVolumeSpecName "kube-api-access-zlhzs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.013706 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.013765 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zlhzs\" (UniqueName: \"kubernetes.io/projected/d5782118-cbff-484d-a3fc-6be69adb37bb-kube-api-access-zlhzs\") on node \"crc\" DevicePath \"\"" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.294929 4923 generic.go:334] "Generic (PLEG): container finished" podID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerID="2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53" exitCode=0 Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.294975 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pf7d9" event={"ID":"d5782118-cbff-484d-a3fc-6be69adb37bb","Type":"ContainerDied","Data":"2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53"} Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.295431 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pf7d9" event={"ID":"d5782118-cbff-484d-a3fc-6be69adb37bb","Type":"ContainerDied","Data":"4958a7c0e176dd14e3f8a3479ea76ca29862c81e6d839f25d5ee7ed458029c4f"} Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.295464 4923 scope.go:117] "RemoveContainer" containerID="2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.295032 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pf7d9" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.323591 4923 scope.go:117] "RemoveContainer" containerID="7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.346105 4923 scope.go:117] "RemoveContainer" containerID="bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.387937 4923 scope.go:117] "RemoveContainer" containerID="2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53" Oct 09 10:53:39 crc kubenswrapper[4923]: E1009 10:53:39.388791 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53\": container with ID starting with 2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53 not found: ID does not exist" containerID="2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.388845 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53"} err="failed to get container status \"2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53\": rpc error: code = NotFound desc = could not find container \"2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53\": container with ID starting with 2e1496b48ede8035dfab5898c6996a553a3bf038a5b29e37c92e046cea4ddf53 not found: ID does not exist" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.388890 4923 scope.go:117] "RemoveContainer" containerID="7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e" Oct 09 10:53:39 crc kubenswrapper[4923]: E1009 10:53:39.389443 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e\": container with ID starting with 7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e not found: ID does not exist" containerID="7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.389501 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e"} err="failed to get container status \"7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e\": rpc error: code = NotFound desc = could not find container \"7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e\": container with ID starting with 7b1eb2f8e90e50c065dd46af17bbef9ec85e90a8460d77b1b4a01a148bcfb87e not found: ID does not exist" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.389541 4923 scope.go:117] "RemoveContainer" containerID="bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02" Oct 09 10:53:39 crc kubenswrapper[4923]: E1009 10:53:39.389948 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02\": container with ID starting with bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02 not found: ID does not exist" containerID="bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.389990 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02"} err="failed to get container status \"bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02\": rpc error: code = NotFound desc = could not find container \"bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02\": container with ID starting with bd465a089ce36419e3df59ae21a7b63c180bf32b3be4deb10a7f3921bfeeac02 not found: ID does not exist" Oct 09 10:53:39 crc kubenswrapper[4923]: I1009 10:53:39.601783 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:53:39 crc kubenswrapper[4923]: E1009 10:53:39.602238 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:53:41 crc kubenswrapper[4923]: I1009 10:53:41.106896 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d5782118-cbff-484d-a3fc-6be69adb37bb" (UID: "d5782118-cbff-484d-a3fc-6be69adb37bb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:53:41 crc kubenswrapper[4923]: I1009 10:53:41.162126 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d5782118-cbff-484d-a3fc-6be69adb37bb-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:53:41 crc kubenswrapper[4923]: I1009 10:53:41.444820 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pf7d9"] Oct 09 10:53:41 crc kubenswrapper[4923]: I1009 10:53:41.455940 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pf7d9"] Oct 09 10:53:42 crc kubenswrapper[4923]: I1009 10:53:42.612928 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5782118-cbff-484d-a3fc-6be69adb37bb" path="/var/lib/kubelet/pods/d5782118-cbff-484d-a3fc-6be69adb37bb/volumes" Oct 09 10:53:51 crc kubenswrapper[4923]: I1009 10:53:51.602265 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:53:51 crc kubenswrapper[4923]: E1009 10:53:51.603408 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:54:03 crc kubenswrapper[4923]: I1009 10:54:03.601392 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:54:03 crc kubenswrapper[4923]: E1009 10:54:03.602486 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:54:14 crc kubenswrapper[4923]: I1009 10:54:14.609128 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:54:14 crc kubenswrapper[4923]: E1009 10:54:14.610658 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:54:27 crc kubenswrapper[4923]: I1009 10:54:27.602216 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:54:27 crc kubenswrapper[4923]: E1009 10:54:27.603401 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:54:39 crc kubenswrapper[4923]: I1009 10:54:39.601990 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:54:39 crc kubenswrapper[4923]: E1009 10:54:39.603267 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:54:50 crc kubenswrapper[4923]: I1009 10:54:50.602417 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:54:50 crc kubenswrapper[4923]: E1009 10:54:50.604002 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:55:05 crc kubenswrapper[4923]: I1009 10:55:05.602023 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:55:05 crc kubenswrapper[4923]: E1009 10:55:05.603277 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:55:16 crc kubenswrapper[4923]: I1009 10:55:16.602631 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:55:16 crc kubenswrapper[4923]: E1009 10:55:16.604167 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:55:31 crc kubenswrapper[4923]: I1009 10:55:31.601779 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:55:31 crc kubenswrapper[4923]: E1009 10:55:31.602788 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:55:42 crc kubenswrapper[4923]: I1009 10:55:42.601537 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:55:42 crc kubenswrapper[4923]: E1009 10:55:42.602533 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:55:57 crc kubenswrapper[4923]: I1009 10:55:57.602591 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:55:57 crc kubenswrapper[4923]: E1009 10:55:57.603765 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:56:08 crc kubenswrapper[4923]: I1009 10:56:08.601866 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:56:08 crc kubenswrapper[4923]: E1009 10:56:08.602904 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:56:19 crc kubenswrapper[4923]: I1009 10:56:19.880400 4923 generic.go:334] "Generic (PLEG): container finished" podID="81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" containerID="cc6c8052c31e87f89c711e596e77b24658d30852d2a5ad081605cbe6853fc576" exitCode=0 Oct 09 10:56:19 crc kubenswrapper[4923]: I1009 10:56:19.880476 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" event={"ID":"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45","Type":"ContainerDied","Data":"cc6c8052c31e87f89c711e596e77b24658d30852d2a5ad081605cbe6853fc576"} Oct 09 10:56:20 crc kubenswrapper[4923]: I1009 10:56:20.601796 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:56:20 crc kubenswrapper[4923]: E1009 10:56:20.602224 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.328177 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.468570 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-secret-0\") pod \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.468705 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ssh-key\") pod \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.468836 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7m9f2\" (UniqueName: \"kubernetes.io/projected/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-kube-api-access-7m9f2\") pod \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.468943 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ceph\") pod \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.468998 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-inventory\") pod \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.469038 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-combined-ca-bundle\") pod \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\" (UID: \"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45\") " Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.475849 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ceph" (OuterVolumeSpecName: "ceph") pod "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" (UID: "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.475930 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-kube-api-access-7m9f2" (OuterVolumeSpecName: "kube-api-access-7m9f2") pod "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" (UID: "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45"). InnerVolumeSpecName "kube-api-access-7m9f2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.476339 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" (UID: "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.500348 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" (UID: "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.501035 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-inventory" (OuterVolumeSpecName: "inventory") pod "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" (UID: "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.504666 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" (UID: "81549d8a-1bfc-4261-b7bc-7edb5b2d9a45"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.571640 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7m9f2\" (UniqueName: \"kubernetes.io/projected/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-kube-api-access-7m9f2\") on node \"crc\" DevicePath \"\"" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.571692 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.571706 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.571717 4923 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.571729 4923 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.571742 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/81549d8a-1bfc-4261-b7bc-7edb5b2d9a45-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.918079 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" event={"ID":"81549d8a-1bfc-4261-b7bc-7edb5b2d9a45","Type":"ContainerDied","Data":"047550eec0873de661d50ab9faef876712212d632c921450f35971751b5dfeff"} Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.918172 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="047550eec0873de661d50ab9faef876712212d632c921450f35971751b5dfeff" Oct 09 10:56:21 crc kubenswrapper[4923]: I1009 10:56:21.918338 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2f26f" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.025829 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h"] Oct 09 10:56:22 crc kubenswrapper[4923]: E1009 10:56:22.027987 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerName="extract-utilities" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.028106 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerName="extract-utilities" Oct 09 10:56:22 crc kubenswrapper[4923]: E1009 10:56:22.028206 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerName="extract-content" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.028275 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerName="extract-content" Oct 09 10:56:22 crc kubenswrapper[4923]: E1009 10:56:22.028332 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerName="registry-server" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.028388 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerName="registry-server" Oct 09 10:56:22 crc kubenswrapper[4923]: E1009 10:56:22.028456 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.028517 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.028765 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5782118-cbff-484d-a3fc-6be69adb37bb" containerName="registry-server" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.028866 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="81549d8a-1bfc-4261-b7bc-7edb5b2d9a45" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.029806 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.032556 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.032580 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.032973 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.033087 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-fzc6p" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.033237 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.033412 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.033588 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.034641 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.035285 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.038349 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h"] Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.081687 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082097 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082169 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082203 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbbxx\" (UniqueName: \"kubernetes.io/projected/c2ee9657-939a-4acf-a25b-eb315ebec083-kube-api-access-lbbxx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082234 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082301 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082324 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082362 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082407 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082438 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.082471 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184134 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184216 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184250 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184308 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184329 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbbxx\" (UniqueName: \"kubernetes.io/projected/c2ee9657-939a-4acf-a25b-eb315ebec083-kube-api-access-lbbxx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184357 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184407 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184428 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184461 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184498 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.184520 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.186717 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.186759 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.189975 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.190962 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.191130 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.191475 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.192037 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.192651 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.192734 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.195718 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.213014 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbbxx\" (UniqueName: \"kubernetes.io/projected/c2ee9657-939a-4acf-a25b-eb315ebec083-kube-api-access-lbbxx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.394301 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.936259 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h"] Oct 09 10:56:22 crc kubenswrapper[4923]: I1009 10:56:22.942979 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 10:56:23 crc kubenswrapper[4923]: I1009 10:56:23.936210 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" event={"ID":"c2ee9657-939a-4acf-a25b-eb315ebec083","Type":"ContainerStarted","Data":"5b3ba4ad13688d595c47ec30d52e30c5acd3c308eb15d392441874a6e10da3f6"} Oct 09 10:56:23 crc kubenswrapper[4923]: I1009 10:56:23.936593 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" event={"ID":"c2ee9657-939a-4acf-a25b-eb315ebec083","Type":"ContainerStarted","Data":"0d99dae8024da65897698c057b13880efb3b53f258b3af01b0ad3258f368f8a0"} Oct 09 10:56:23 crc kubenswrapper[4923]: I1009 10:56:23.966327 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" podStartSLOduration=1.287443803 podStartE2EDuration="1.966293505s" podCreationTimestamp="2025-10-09 10:56:22 +0000 UTC" firstStartedPulling="2025-10-09 10:56:22.942748946 +0000 UTC m=+3069.010930702" lastFinishedPulling="2025-10-09 10:56:23.621598648 +0000 UTC m=+3069.689780404" observedRunningTime="2025-10-09 10:56:23.960224207 +0000 UTC m=+3070.028405963" watchObservedRunningTime="2025-10-09 10:56:23.966293505 +0000 UTC m=+3070.034475271" Oct 09 10:56:31 crc kubenswrapper[4923]: I1009 10:56:31.601736 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:56:31 crc kubenswrapper[4923]: E1009 10:56:31.602500 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:56:43 crc kubenswrapper[4923]: I1009 10:56:43.601893 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:56:43 crc kubenswrapper[4923]: E1009 10:56:43.603072 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.537929 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-drjvp"] Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.541422 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.557355 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-drjvp"] Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.602513 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:56:58 crc kubenswrapper[4923]: E1009 10:56:58.602933 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.696671 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmwcg\" (UniqueName: \"kubernetes.io/projected/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-kube-api-access-tmwcg\") pod \"certified-operators-drjvp\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.696979 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-catalog-content\") pod \"certified-operators-drjvp\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.697239 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-utilities\") pod \"certified-operators-drjvp\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.800159 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-utilities\") pod \"certified-operators-drjvp\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.800252 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmwcg\" (UniqueName: \"kubernetes.io/projected/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-kube-api-access-tmwcg\") pod \"certified-operators-drjvp\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.800327 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-catalog-content\") pod \"certified-operators-drjvp\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.800662 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-utilities\") pod \"certified-operators-drjvp\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.800883 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-catalog-content\") pod \"certified-operators-drjvp\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.824657 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmwcg\" (UniqueName: \"kubernetes.io/projected/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-kube-api-access-tmwcg\") pod \"certified-operators-drjvp\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:58 crc kubenswrapper[4923]: I1009 10:56:58.870545 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:56:59 crc kubenswrapper[4923]: I1009 10:56:59.436359 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-drjvp"] Oct 09 10:57:00 crc kubenswrapper[4923]: I1009 10:57:00.276713 4923 generic.go:334] "Generic (PLEG): container finished" podID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerID="e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b" exitCode=0 Oct 09 10:57:00 crc kubenswrapper[4923]: I1009 10:57:00.276856 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drjvp" event={"ID":"5fe2ece0-2a79-48c7-b5da-4e39acd503ad","Type":"ContainerDied","Data":"e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b"} Oct 09 10:57:00 crc kubenswrapper[4923]: I1009 10:57:00.277304 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drjvp" event={"ID":"5fe2ece0-2a79-48c7-b5da-4e39acd503ad","Type":"ContainerStarted","Data":"86f391d910503347d64fa8e2ca3ed12900da0b39525aa6c741fde5a28b4fc319"} Oct 09 10:57:01 crc kubenswrapper[4923]: I1009 10:57:01.288092 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drjvp" event={"ID":"5fe2ece0-2a79-48c7-b5da-4e39acd503ad","Type":"ContainerStarted","Data":"783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96"} Oct 09 10:57:02 crc kubenswrapper[4923]: I1009 10:57:02.302136 4923 generic.go:334] "Generic (PLEG): container finished" podID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerID="783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96" exitCode=0 Oct 09 10:57:02 crc kubenswrapper[4923]: I1009 10:57:02.302195 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drjvp" event={"ID":"5fe2ece0-2a79-48c7-b5da-4e39acd503ad","Type":"ContainerDied","Data":"783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96"} Oct 09 10:57:03 crc kubenswrapper[4923]: I1009 10:57:03.325174 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drjvp" event={"ID":"5fe2ece0-2a79-48c7-b5da-4e39acd503ad","Type":"ContainerStarted","Data":"8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088"} Oct 09 10:57:03 crc kubenswrapper[4923]: I1009 10:57:03.354412 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-drjvp" podStartSLOduration=2.806935469 podStartE2EDuration="5.35438951s" podCreationTimestamp="2025-10-09 10:56:58 +0000 UTC" firstStartedPulling="2025-10-09 10:57:00.282264262 +0000 UTC m=+3106.350446018" lastFinishedPulling="2025-10-09 10:57:02.829718303 +0000 UTC m=+3108.897900059" observedRunningTime="2025-10-09 10:57:03.350599826 +0000 UTC m=+3109.418781612" watchObservedRunningTime="2025-10-09 10:57:03.35438951 +0000 UTC m=+3109.422571266" Oct 09 10:57:08 crc kubenswrapper[4923]: I1009 10:57:08.872630 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:57:08 crc kubenswrapper[4923]: I1009 10:57:08.873492 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:57:08 crc kubenswrapper[4923]: I1009 10:57:08.941538 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:57:09 crc kubenswrapper[4923]: I1009 10:57:09.441976 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:57:09 crc kubenswrapper[4923]: I1009 10:57:09.601541 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:57:09 crc kubenswrapper[4923]: E1009 10:57:09.601989 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:57:10 crc kubenswrapper[4923]: I1009 10:57:10.925025 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-drjvp"] Oct 09 10:57:11 crc kubenswrapper[4923]: I1009 10:57:11.409334 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-drjvp" podUID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerName="registry-server" containerID="cri-o://8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088" gracePeriod=2 Oct 09 10:57:11 crc kubenswrapper[4923]: I1009 10:57:11.850065 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.000824 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-utilities\") pod \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.001150 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmwcg\" (UniqueName: \"kubernetes.io/projected/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-kube-api-access-tmwcg\") pod \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.001313 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-catalog-content\") pod \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\" (UID: \"5fe2ece0-2a79-48c7-b5da-4e39acd503ad\") " Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.001876 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-utilities" (OuterVolumeSpecName: "utilities") pod "5fe2ece0-2a79-48c7-b5da-4e39acd503ad" (UID: "5fe2ece0-2a79-48c7-b5da-4e39acd503ad"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.002123 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.008111 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-kube-api-access-tmwcg" (OuterVolumeSpecName: "kube-api-access-tmwcg") pod "5fe2ece0-2a79-48c7-b5da-4e39acd503ad" (UID: "5fe2ece0-2a79-48c7-b5da-4e39acd503ad"). InnerVolumeSpecName "kube-api-access-tmwcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.075715 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fe2ece0-2a79-48c7-b5da-4e39acd503ad" (UID: "5fe2ece0-2a79-48c7-b5da-4e39acd503ad"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.104356 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmwcg\" (UniqueName: \"kubernetes.io/projected/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-kube-api-access-tmwcg\") on node \"crc\" DevicePath \"\"" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.104407 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fe2ece0-2a79-48c7-b5da-4e39acd503ad-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.421955 4923 generic.go:334] "Generic (PLEG): container finished" podID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerID="8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088" exitCode=0 Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.422014 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drjvp" event={"ID":"5fe2ece0-2a79-48c7-b5da-4e39acd503ad","Type":"ContainerDied","Data":"8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088"} Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.422351 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-drjvp" event={"ID":"5fe2ece0-2a79-48c7-b5da-4e39acd503ad","Type":"ContainerDied","Data":"86f391d910503347d64fa8e2ca3ed12900da0b39525aa6c741fde5a28b4fc319"} Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.422384 4923 scope.go:117] "RemoveContainer" containerID="8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.422097 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-drjvp" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.445365 4923 scope.go:117] "RemoveContainer" containerID="783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.467582 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-drjvp"] Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.474786 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-drjvp"] Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.491263 4923 scope.go:117] "RemoveContainer" containerID="e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.528317 4923 scope.go:117] "RemoveContainer" containerID="8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088" Oct 09 10:57:12 crc kubenswrapper[4923]: E1009 10:57:12.528950 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088\": container with ID starting with 8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088 not found: ID does not exist" containerID="8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.529016 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088"} err="failed to get container status \"8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088\": rpc error: code = NotFound desc = could not find container \"8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088\": container with ID starting with 8793e87fd8a83097a6c596927dfe6443803d7e1241e6faff705aa0a7edd92088 not found: ID does not exist" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.529044 4923 scope.go:117] "RemoveContainer" containerID="783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96" Oct 09 10:57:12 crc kubenswrapper[4923]: E1009 10:57:12.529390 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96\": container with ID starting with 783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96 not found: ID does not exist" containerID="783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.529419 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96"} err="failed to get container status \"783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96\": rpc error: code = NotFound desc = could not find container \"783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96\": container with ID starting with 783118f2fa4764896c9230fd1f707abbcfa9abe9defc67f0e24f759f54e6ca96 not found: ID does not exist" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.529435 4923 scope.go:117] "RemoveContainer" containerID="e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b" Oct 09 10:57:12 crc kubenswrapper[4923]: E1009 10:57:12.529961 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b\": container with ID starting with e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b not found: ID does not exist" containerID="e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.530018 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b"} err="failed to get container status \"e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b\": rpc error: code = NotFound desc = could not find container \"e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b\": container with ID starting with e25028c03153f8d2285654af635cd17cbfff8b440a7577e47ff5ceeaefa7df8b not found: ID does not exist" Oct 09 10:57:12 crc kubenswrapper[4923]: I1009 10:57:12.616176 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" path="/var/lib/kubelet/pods/5fe2ece0-2a79-48c7-b5da-4e39acd503ad/volumes" Oct 09 10:57:22 crc kubenswrapper[4923]: I1009 10:57:22.602381 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:57:22 crc kubenswrapper[4923]: E1009 10:57:22.603625 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:57:33 crc kubenswrapper[4923]: I1009 10:57:33.602177 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:57:33 crc kubenswrapper[4923]: E1009 10:57:33.603546 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:57:44 crc kubenswrapper[4923]: I1009 10:57:44.607662 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:57:44 crc kubenswrapper[4923]: E1009 10:57:44.609045 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:57:55 crc kubenswrapper[4923]: I1009 10:57:55.604308 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:57:55 crc kubenswrapper[4923]: E1009 10:57:55.606859 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:58:10 crc kubenswrapper[4923]: I1009 10:58:10.601886 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:58:10 crc kubenswrapper[4923]: E1009 10:58:10.603147 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:58:21 crc kubenswrapper[4923]: I1009 10:58:21.601495 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:58:21 crc kubenswrapper[4923]: E1009 10:58:21.602574 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 10:58:33 crc kubenswrapper[4923]: I1009 10:58:33.602108 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 10:58:34 crc kubenswrapper[4923]: I1009 10:58:34.222049 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"0c284c9aa97a8197efff9350f0a239489485f6b80d76c31d68671f5e5bb2b8a3"} Oct 09 10:59:59 crc kubenswrapper[4923]: I1009 10:59:59.086559 4923 generic.go:334] "Generic (PLEG): container finished" podID="c2ee9657-939a-4acf-a25b-eb315ebec083" containerID="5b3ba4ad13688d595c47ec30d52e30c5acd3c308eb15d392441874a6e10da3f6" exitCode=0 Oct 09 10:59:59 crc kubenswrapper[4923]: I1009 10:59:59.086673 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" event={"ID":"c2ee9657-939a-4acf-a25b-eb315ebec083","Type":"ContainerDied","Data":"5b3ba4ad13688d595c47ec30d52e30c5acd3c308eb15d392441874a6e10da3f6"} Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.176305 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l"] Oct 09 11:00:00 crc kubenswrapper[4923]: E1009 11:00:00.178204 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerName="extract-content" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.178249 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerName="extract-content" Oct 09 11:00:00 crc kubenswrapper[4923]: E1009 11:00:00.178292 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerName="registry-server" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.178303 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerName="registry-server" Oct 09 11:00:00 crc kubenswrapper[4923]: E1009 11:00:00.178315 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerName="extract-utilities" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.178323 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerName="extract-utilities" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.178583 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fe2ece0-2a79-48c7-b5da-4e39acd503ad" containerName="registry-server" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.179602 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.186314 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.189055 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.196396 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l"] Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.327488 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fe0d9d8f-2b10-4800-a738-b38c18b40cca-secret-volume\") pod \"collect-profiles-29333460-j592l\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.327570 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cdsm\" (UniqueName: \"kubernetes.io/projected/fe0d9d8f-2b10-4800-a738-b38c18b40cca-kube-api-access-2cdsm\") pod \"collect-profiles-29333460-j592l\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.327726 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fe0d9d8f-2b10-4800-a738-b38c18b40cca-config-volume\") pod \"collect-profiles-29333460-j592l\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.429967 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fe0d9d8f-2b10-4800-a738-b38c18b40cca-config-volume\") pod \"collect-profiles-29333460-j592l\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.430075 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fe0d9d8f-2b10-4800-a738-b38c18b40cca-secret-volume\") pod \"collect-profiles-29333460-j592l\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.430129 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cdsm\" (UniqueName: \"kubernetes.io/projected/fe0d9d8f-2b10-4800-a738-b38c18b40cca-kube-api-access-2cdsm\") pod \"collect-profiles-29333460-j592l\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.433173 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fe0d9d8f-2b10-4800-a738-b38c18b40cca-config-volume\") pod \"collect-profiles-29333460-j592l\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.458940 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fe0d9d8f-2b10-4800-a738-b38c18b40cca-secret-volume\") pod \"collect-profiles-29333460-j592l\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.486311 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cdsm\" (UniqueName: \"kubernetes.io/projected/fe0d9d8f-2b10-4800-a738-b38c18b40cca-kube-api-access-2cdsm\") pod \"collect-profiles-29333460-j592l\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.520396 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.818660 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.950690 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-custom-ceph-combined-ca-bundle\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.950937 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph-nova-0\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.950983 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ssh-key\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.951050 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-inventory\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.951114 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-0\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.951253 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-extra-config-0\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.951282 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbbxx\" (UniqueName: \"kubernetes.io/projected/c2ee9657-939a-4acf-a25b-eb315ebec083-kube-api-access-lbbxx\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.951308 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-1\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.951660 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.951945 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-0\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.951980 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-1\") pod \"c2ee9657-939a-4acf-a25b-eb315ebec083\" (UID: \"c2ee9657-939a-4acf-a25b-eb315ebec083\") " Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.958565 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.974151 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph" (OuterVolumeSpecName: "ceph") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.997385 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2ee9657-939a-4acf-a25b-eb315ebec083-kube-api-access-lbbxx" (OuterVolumeSpecName: "kube-api-access-lbbxx") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "kube-api-access-lbbxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:00:00 crc kubenswrapper[4923]: I1009 11:00:00.999843 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-inventory" (OuterVolumeSpecName: "inventory") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.001726 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.023041 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.047522 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.055261 4923 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.055345 4923 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.055360 4923 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.055373 4923 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.055390 4923 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.055424 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbbxx\" (UniqueName: \"kubernetes.io/projected/c2ee9657-939a-4acf-a25b-eb315ebec083-kube-api-access-lbbxx\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.055439 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.057105 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.061234 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l"] Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.078024 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.084801 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.086378 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "c2ee9657-939a-4acf-a25b-eb315ebec083" (UID: "c2ee9657-939a-4acf-a25b-eb315ebec083"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.112659 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" event={"ID":"c2ee9657-939a-4acf-a25b-eb315ebec083","Type":"ContainerDied","Data":"0d99dae8024da65897698c057b13880efb3b53f258b3af01b0ad3258f368f8a0"} Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.112723 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d99dae8024da65897698c057b13880efb3b53f258b3af01b0ad3258f368f8a0" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.113376 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.114700 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" event={"ID":"fe0d9d8f-2b10-4800-a738-b38c18b40cca","Type":"ContainerStarted","Data":"856c759956b4104d5d407a9252d542703a624383eb2a1f993e65f7639ca7836f"} Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.160679 4923 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.160738 4923 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.160771 4923 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:01 crc kubenswrapper[4923]: I1009 11:00:01.160792 4923 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2ee9657-939a-4acf-a25b-eb315ebec083-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:02 crc kubenswrapper[4923]: I1009 11:00:02.129270 4923 generic.go:334] "Generic (PLEG): container finished" podID="fe0d9d8f-2b10-4800-a738-b38c18b40cca" containerID="166c9630f58fc398227ded4aaa01f43bc308eb8a2a8101c50c68d99513a54324" exitCode=0 Oct 09 11:00:02 crc kubenswrapper[4923]: I1009 11:00:02.129706 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" event={"ID":"fe0d9d8f-2b10-4800-a738-b38c18b40cca","Type":"ContainerDied","Data":"166c9630f58fc398227ded4aaa01f43bc308eb8a2a8101c50c68d99513a54324"} Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.518585 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.611716 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cdsm\" (UniqueName: \"kubernetes.io/projected/fe0d9d8f-2b10-4800-a738-b38c18b40cca-kube-api-access-2cdsm\") pod \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.612295 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fe0d9d8f-2b10-4800-a738-b38c18b40cca-config-volume\") pod \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.612509 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fe0d9d8f-2b10-4800-a738-b38c18b40cca-secret-volume\") pod \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\" (UID: \"fe0d9d8f-2b10-4800-a738-b38c18b40cca\") " Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.613133 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe0d9d8f-2b10-4800-a738-b38c18b40cca-config-volume" (OuterVolumeSpecName: "config-volume") pod "fe0d9d8f-2b10-4800-a738-b38c18b40cca" (UID: "fe0d9d8f-2b10-4800-a738-b38c18b40cca"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.615405 4923 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/fe0d9d8f-2b10-4800-a738-b38c18b40cca-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.621068 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe0d9d8f-2b10-4800-a738-b38c18b40cca-kube-api-access-2cdsm" (OuterVolumeSpecName: "kube-api-access-2cdsm") pod "fe0d9d8f-2b10-4800-a738-b38c18b40cca" (UID: "fe0d9d8f-2b10-4800-a738-b38c18b40cca"). InnerVolumeSpecName "kube-api-access-2cdsm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.621122 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe0d9d8f-2b10-4800-a738-b38c18b40cca-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "fe0d9d8f-2b10-4800-a738-b38c18b40cca" (UID: "fe0d9d8f-2b10-4800-a738-b38c18b40cca"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.717108 4923 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/fe0d9d8f-2b10-4800-a738-b38c18b40cca-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:03 crc kubenswrapper[4923]: I1009 11:00:03.717166 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cdsm\" (UniqueName: \"kubernetes.io/projected/fe0d9d8f-2b10-4800-a738-b38c18b40cca-kube-api-access-2cdsm\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:04 crc kubenswrapper[4923]: I1009 11:00:04.149783 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" event={"ID":"fe0d9d8f-2b10-4800-a738-b38c18b40cca","Type":"ContainerDied","Data":"856c759956b4104d5d407a9252d542703a624383eb2a1f993e65f7639ca7836f"} Oct 09 11:00:04 crc kubenswrapper[4923]: I1009 11:00:04.149847 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="856c759956b4104d5d407a9252d542703a624383eb2a1f993e65f7639ca7836f" Oct 09 11:00:04 crc kubenswrapper[4923]: I1009 11:00:04.149873 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333460-j592l" Oct 09 11:00:04 crc kubenswrapper[4923]: I1009 11:00:04.633877 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9"] Oct 09 11:00:04 crc kubenswrapper[4923]: I1009 11:00:04.651831 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333415-h4fp9"] Oct 09 11:00:06 crc kubenswrapper[4923]: I1009 11:00:06.614468 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09667cd2-7829-459b-a8bb-0cd255c221ab" path="/var/lib/kubelet/pods/09667cd2-7829-459b-a8bb-0cd255c221ab/volumes" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.326317 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 09 11:00:17 crc kubenswrapper[4923]: E1009 11:00:17.327589 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe0d9d8f-2b10-4800-a738-b38c18b40cca" containerName="collect-profiles" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.327608 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe0d9d8f-2b10-4800-a738-b38c18b40cca" containerName="collect-profiles" Oct 09 11:00:17 crc kubenswrapper[4923]: E1009 11:00:17.327636 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2ee9657-939a-4acf-a25b-eb315ebec083" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.327645 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2ee9657-939a-4acf-a25b-eb315ebec083" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.327860 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2ee9657-939a-4acf-a25b-eb315ebec083" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.327871 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe0d9d8f-2b10-4800-a738-b38c18b40cca" containerName="collect-profiles" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.329111 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.335215 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.338866 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.346644 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.428132 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.428227 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.428272 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.428302 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.428380 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52lct\" (UniqueName: \"kubernetes.io/projected/23f48866-c443-458e-9a5a-dd2243b86249-kube-api-access-52lct\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.428509 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.428996 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.429183 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-sys\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.429244 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.429464 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-dev\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.429525 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-run\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.429602 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.429704 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.429808 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/23f48866-c443-458e-9a5a-dd2243b86249-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.429865 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.429956 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.437708 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.439776 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.444273 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.459613 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531438 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531497 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/23f48866-c443-458e-9a5a-dd2243b86249-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531530 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-lib-modules\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531561 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffvdt\" (UniqueName: \"kubernetes.io/projected/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-kube-api-access-ffvdt\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531585 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531610 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531633 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531644 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531692 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531741 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-config-data\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531867 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-sys\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531917 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531938 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531955 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.531963 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532038 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532087 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532114 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52lct\" (UniqueName: \"kubernetes.io/projected/23f48866-c443-458e-9a5a-dd2243b86249-kube-api-access-52lct\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532156 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532261 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532261 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532303 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-dev\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532347 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-scripts\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532382 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532408 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532430 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532446 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-ceph\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532485 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-sys\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532509 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532534 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-run\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532555 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-dev\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532574 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-run\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532596 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532615 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532633 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532670 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532789 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-sys\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532846 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532860 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532895 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-run\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532908 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-dev\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.532983 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/23f48866-c443-458e-9a5a-dd2243b86249-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.540056 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.541445 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.542164 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.542232 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/23f48866-c443-458e-9a5a-dd2243b86249-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.553547 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/23f48866-c443-458e-9a5a-dd2243b86249-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.553848 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52lct\" (UniqueName: \"kubernetes.io/projected/23f48866-c443-458e-9a5a-dd2243b86249-kube-api-access-52lct\") pod \"cinder-volume-volume1-0\" (UID: \"23f48866-c443-458e-9a5a-dd2243b86249\") " pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635168 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffvdt\" (UniqueName: \"kubernetes.io/projected/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-kube-api-access-ffvdt\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635258 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635316 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-config-data\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635338 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-sys\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635405 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635446 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635470 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635471 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-sys\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635497 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-dev\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635536 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635613 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-scripts\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635656 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-dev\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635720 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635671 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635812 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635843 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-ceph\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635931 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-run\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635974 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.636011 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.636014 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-run\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.635932 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-etc-nvme\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.636048 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.636058 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.636112 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.636140 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-lib-modules\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.636256 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-lib-modules\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.640903 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-config-data-custom\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.640941 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.641009 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-ceph\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.644454 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-scripts\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.645500 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-config-data\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.658998 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.660437 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffvdt\" (UniqueName: \"kubernetes.io/projected/5d650442-1c4b-48cd-8b94-e9dd6e51cae3-kube-api-access-ffvdt\") pod \"cinder-backup-0\" (UID: \"5d650442-1c4b-48cd-8b94-e9dd6e51cae3\") " pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.760114 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.934818 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-mxv9k"] Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.938176 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mxv9k" Oct 09 11:00:17 crc kubenswrapper[4923]: I1009 11:00:17.947227 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mxv9k"] Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.047333 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7cp2\" (UniqueName: \"kubernetes.io/projected/f093ece7-e8f5-4beb-8f89-612afece0428-kube-api-access-m7cp2\") pod \"manila-db-create-mxv9k\" (UID: \"f093ece7-e8f5-4beb-8f89-612afece0428\") " pod="openstack/manila-db-create-mxv9k" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.110407 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.112445 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.115676 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.115979 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-mnjv6" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.116173 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.118057 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.133326 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.151070 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7cp2\" (UniqueName: \"kubernetes.io/projected/f093ece7-e8f5-4beb-8f89-612afece0428-kube-api-access-m7cp2\") pod \"manila-db-create-mxv9k\" (UID: \"f093ece7-e8f5-4beb-8f89-612afece0428\") " pod="openstack/manila-db-create-mxv9k" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.197106 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7cp2\" (UniqueName: \"kubernetes.io/projected/f093ece7-e8f5-4beb-8f89-612afece0428-kube-api-access-m7cp2\") pod \"manila-db-create-mxv9k\" (UID: \"f093ece7-e8f5-4beb-8f89-612afece0428\") " pod="openstack/manila-db-create-mxv9k" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.250741 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.253030 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-config-data\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.253102 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-scripts\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.253163 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-ceph\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.253225 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.253476 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.253514 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7krgm\" (UniqueName: \"kubernetes.io/projected/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-kube-api-access-7krgm\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.253577 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-logs\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.253615 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.253641 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.254719 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.257372 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.257474 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.261887 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.270440 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.287380 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"23f48866-c443-458e-9a5a-dd2243b86249","Type":"ContainerStarted","Data":"8629658b4d280b474c1a1b54bf3fed2fe54cae535d5b9d90c921b4a26db6b448"} Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.290731 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mxv9k" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356137 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356206 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7krgm\" (UniqueName: \"kubernetes.io/projected/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-kube-api-access-7krgm\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356267 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-logs\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356307 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356331 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356355 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356414 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356464 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356500 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-config-data\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356531 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356567 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-ceph\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356589 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-scripts\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356619 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spxrx\" (UniqueName: \"kubernetes.io/projected/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-kube-api-access-spxrx\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356644 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356689 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-logs\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356718 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-ceph\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356743 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.356805 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.357225 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.358839 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.359226 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-logs\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.360844 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-config-data\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.363553 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-scripts\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.363704 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.368946 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-ceph\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.369511 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.381844 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7krgm\" (UniqueName: \"kubernetes.io/projected/5a6db3b3-10aa-4af5-958a-15f9ca7f5102-kube-api-access-7krgm\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.400912 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-0\" (UID: \"5a6db3b3-10aa-4af5-958a-15f9ca7f5102\") " pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.434334 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.447684 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Oct 09 11:00:18 crc kubenswrapper[4923]: W1009 11:00:18.455800 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d650442_1c4b_48cd_8b94_e9dd6e51cae3.slice/crio-5d48c1cd9efbde517d1dd3be703ba96e8168eadf37e2746e90d17578eede626a WatchSource:0}: Error finding container 5d48c1cd9efbde517d1dd3be703ba96e8168eadf37e2746e90d17578eede626a: Status 404 returned error can't find the container with id 5d48c1cd9efbde517d1dd3be703ba96e8168eadf37e2746e90d17578eede626a Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.458937 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spxrx\" (UniqueName: \"kubernetes.io/projected/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-kube-api-access-spxrx\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.458989 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.459042 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-logs\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.459074 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.459867 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.459937 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.459992 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.459861 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-logs\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.460268 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.460398 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-ceph\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.460597 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.461001 4923 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.463855 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.464287 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.465571 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-ceph\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.470389 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.479549 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.486338 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spxrx\" (UniqueName: \"kubernetes.io/projected/9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5-kube-api-access-spxrx\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.505991 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5\") " pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.578183 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:18 crc kubenswrapper[4923]: I1009 11:00:18.794331 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-mxv9k"] Oct 09 11:00:19 crc kubenswrapper[4923]: I1009 11:00:19.031575 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 11:00:19 crc kubenswrapper[4923]: W1009 11:00:19.051695 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a6db3b3_10aa_4af5_958a_15f9ca7f5102.slice/crio-e232202837318a6a0d2242e4ce3171e9ff66ac5c0527668e9135623df99fa184 WatchSource:0}: Error finding container e232202837318a6a0d2242e4ce3171e9ff66ac5c0527668e9135623df99fa184: Status 404 returned error can't find the container with id e232202837318a6a0d2242e4ce3171e9ff66ac5c0527668e9135623df99fa184 Oct 09 11:00:19 crc kubenswrapper[4923]: I1009 11:00:19.208806 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 11:00:19 crc kubenswrapper[4923]: I1009 11:00:19.306591 4923 generic.go:334] "Generic (PLEG): container finished" podID="f093ece7-e8f5-4beb-8f89-612afece0428" containerID="201576e47d13293de82c7debb17e5b304dc5a56b525ae1208a07029a1b550437" exitCode=0 Oct 09 11:00:19 crc kubenswrapper[4923]: I1009 11:00:19.306717 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mxv9k" event={"ID":"f093ece7-e8f5-4beb-8f89-612afece0428","Type":"ContainerDied","Data":"201576e47d13293de82c7debb17e5b304dc5a56b525ae1208a07029a1b550437"} Oct 09 11:00:19 crc kubenswrapper[4923]: I1009 11:00:19.306812 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mxv9k" event={"ID":"f093ece7-e8f5-4beb-8f89-612afece0428","Type":"ContainerStarted","Data":"eda10b416d6d8ce6083e38516615e14e8785e71dbc6ff3826736f42d9115c385"} Oct 09 11:00:19 crc kubenswrapper[4923]: I1009 11:00:19.309815 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5a6db3b3-10aa-4af5-958a-15f9ca7f5102","Type":"ContainerStarted","Data":"e232202837318a6a0d2242e4ce3171e9ff66ac5c0527668e9135623df99fa184"} Oct 09 11:00:19 crc kubenswrapper[4923]: I1009 11:00:19.312326 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5d650442-1c4b-48cd-8b94-e9dd6e51cae3","Type":"ContainerStarted","Data":"5d48c1cd9efbde517d1dd3be703ba96e8168eadf37e2746e90d17578eede626a"} Oct 09 11:00:19 crc kubenswrapper[4923]: I1009 11:00:19.314275 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5","Type":"ContainerStarted","Data":"82efaa26072d35cb02f80a452517b6aa58b1b92399ad17122fdd3b5ea75dba5a"} Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.270381 4923 scope.go:117] "RemoveContainer" containerID="093c9f9c8a360e261fdfe3790d91daac15d0d48833a1881979182167619b3917" Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.328947 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5","Type":"ContainerStarted","Data":"16936217a2863ef1676dbff9dbe331efd02c48074f2424fb597f2b55908cdfd9"} Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.332291 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"23f48866-c443-458e-9a5a-dd2243b86249","Type":"ContainerStarted","Data":"38fa0d85e30fa7ebbed2435553116c252db0149183a8c5139037dc9f12bcea82"} Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.332353 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"23f48866-c443-458e-9a5a-dd2243b86249","Type":"ContainerStarted","Data":"febdfb3641d01f7d38b8eb5eb43cb181c3c892f0a430912d3df42d0cb143da1a"} Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.336169 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5a6db3b3-10aa-4af5-958a-15f9ca7f5102","Type":"ContainerStarted","Data":"4bdc7c22e59f30f900006b6435bd7f7625865a7a74b86439ff7f8dbcee8054f7"} Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.366483 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5d650442-1c4b-48cd-8b94-e9dd6e51cae3","Type":"ContainerStarted","Data":"3b3a0739e16be11247a9ce96dd0a277b6395ee85d32f7fc43db40e385cb414d1"} Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.366824 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"5d650442-1c4b-48cd-8b94-e9dd6e51cae3","Type":"ContainerStarted","Data":"4ed57feeee97143a836cdaf1e2ca1553ce75538311efccc59d4ed099e898a76e"} Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.371394 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.473292619 podStartE2EDuration="3.371367619s" podCreationTimestamp="2025-10-09 11:00:17 +0000 UTC" firstStartedPulling="2025-10-09 11:00:18.214504767 +0000 UTC m=+3304.282686523" lastFinishedPulling="2025-10-09 11:00:19.112579767 +0000 UTC m=+3305.180761523" observedRunningTime="2025-10-09 11:00:20.363413169 +0000 UTC m=+3306.431595035" watchObservedRunningTime="2025-10-09 11:00:20.371367619 +0000 UTC m=+3306.439549385" Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.406379 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=2.450393603 podStartE2EDuration="3.406340738s" podCreationTimestamp="2025-10-09 11:00:17 +0000 UTC" firstStartedPulling="2025-10-09 11:00:18.469647971 +0000 UTC m=+3304.537829727" lastFinishedPulling="2025-10-09 11:00:19.425595106 +0000 UTC m=+3305.493776862" observedRunningTime="2025-10-09 11:00:20.39448519 +0000 UTC m=+3306.462666976" watchObservedRunningTime="2025-10-09 11:00:20.406340738 +0000 UTC m=+3306.474522494" Oct 09 11:00:20 crc kubenswrapper[4923]: I1009 11:00:20.917369 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mxv9k" Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.060281 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7cp2\" (UniqueName: \"kubernetes.io/projected/f093ece7-e8f5-4beb-8f89-612afece0428-kube-api-access-m7cp2\") pod \"f093ece7-e8f5-4beb-8f89-612afece0428\" (UID: \"f093ece7-e8f5-4beb-8f89-612afece0428\") " Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.069110 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f093ece7-e8f5-4beb-8f89-612afece0428-kube-api-access-m7cp2" (OuterVolumeSpecName: "kube-api-access-m7cp2") pod "f093ece7-e8f5-4beb-8f89-612afece0428" (UID: "f093ece7-e8f5-4beb-8f89-612afece0428"). InnerVolumeSpecName "kube-api-access-m7cp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.165136 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7cp2\" (UniqueName: \"kubernetes.io/projected/f093ece7-e8f5-4beb-8f89-612afece0428-kube-api-access-m7cp2\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.378807 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-mxv9k" event={"ID":"f093ece7-e8f5-4beb-8f89-612afece0428","Type":"ContainerDied","Data":"eda10b416d6d8ce6083e38516615e14e8785e71dbc6ff3826736f42d9115c385"} Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.379275 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eda10b416d6d8ce6083e38516615e14e8785e71dbc6ff3826736f42d9115c385" Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.378818 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-mxv9k" Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.382848 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"5a6db3b3-10aa-4af5-958a-15f9ca7f5102","Type":"ContainerStarted","Data":"79554982f6a5d7e10ee912eedb8b0b0d97905c5b780d2df2ba25e93d50a8f2bb"} Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.393296 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5","Type":"ContainerStarted","Data":"98103af9d424f0122e19fc1419b612479e4c2314bfed48fcb1c045c01329dc0d"} Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.421347 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.42131269 podStartE2EDuration="4.42131269s" podCreationTimestamp="2025-10-09 11:00:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 11:00:21.410236202 +0000 UTC m=+3307.478417958" watchObservedRunningTime="2025-10-09 11:00:21.42131269 +0000 UTC m=+3307.489494466" Oct 09 11:00:21 crc kubenswrapper[4923]: I1009 11:00:21.457340 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.457305138 podStartE2EDuration="4.457305138s" podCreationTimestamp="2025-10-09 11:00:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 11:00:21.440877442 +0000 UTC m=+3307.509059208" watchObservedRunningTime="2025-10-09 11:00:21.457305138 +0000 UTC m=+3307.525486884" Oct 09 11:00:22 crc kubenswrapper[4923]: I1009 11:00:22.659585 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:22 crc kubenswrapper[4923]: I1009 11:00:22.761094 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Oct 09 11:00:27 crc kubenswrapper[4923]: I1009 11:00:27.910434 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.060787 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-37f2-account-create-64wqp"] Oct 09 11:00:28 crc kubenswrapper[4923]: E1009 11:00:28.061951 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f093ece7-e8f5-4beb-8f89-612afece0428" containerName="mariadb-database-create" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.061977 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f093ece7-e8f5-4beb-8f89-612afece0428" containerName="mariadb-database-create" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.062232 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f093ece7-e8f5-4beb-8f89-612afece0428" containerName="mariadb-database-create" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.063115 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-37f2-account-create-64wqp" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.066672 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.080045 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-37f2-account-create-64wqp"] Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.118921 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.126216 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vjssk\" (UniqueName: \"kubernetes.io/projected/53939d29-b8dc-4a59-a7b7-37a8675ced9a-kube-api-access-vjssk\") pod \"manila-37f2-account-create-64wqp\" (UID: \"53939d29-b8dc-4a59-a7b7-37a8675ced9a\") " pod="openstack/manila-37f2-account-create-64wqp" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.229334 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vjssk\" (UniqueName: \"kubernetes.io/projected/53939d29-b8dc-4a59-a7b7-37a8675ced9a-kube-api-access-vjssk\") pod \"manila-37f2-account-create-64wqp\" (UID: \"53939d29-b8dc-4a59-a7b7-37a8675ced9a\") " pod="openstack/manila-37f2-account-create-64wqp" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.263955 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vjssk\" (UniqueName: \"kubernetes.io/projected/53939d29-b8dc-4a59-a7b7-37a8675ced9a-kube-api-access-vjssk\") pod \"manila-37f2-account-create-64wqp\" (UID: \"53939d29-b8dc-4a59-a7b7-37a8675ced9a\") " pod="openstack/manila-37f2-account-create-64wqp" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.399179 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-37f2-account-create-64wqp" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.435195 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.435255 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.503131 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.503569 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.581440 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.584251 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.596470 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.684544 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:28 crc kubenswrapper[4923]: I1009 11:00:28.686415 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:29 crc kubenswrapper[4923]: I1009 11:00:29.015176 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-37f2-account-create-64wqp"] Oct 09 11:00:29 crc kubenswrapper[4923]: I1009 11:00:29.489449 4923 generic.go:334] "Generic (PLEG): container finished" podID="53939d29-b8dc-4a59-a7b7-37a8675ced9a" containerID="3631d6f320264077077990d367e92e333d2e99d4728f0911dfa11666d3ca8a2a" exitCode=0 Oct 09 11:00:29 crc kubenswrapper[4923]: I1009 11:00:29.489903 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-37f2-account-create-64wqp" event={"ID":"53939d29-b8dc-4a59-a7b7-37a8675ced9a","Type":"ContainerDied","Data":"3631d6f320264077077990d367e92e333d2e99d4728f0911dfa11666d3ca8a2a"} Oct 09 11:00:29 crc kubenswrapper[4923]: I1009 11:00:29.489986 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-37f2-account-create-64wqp" event={"ID":"53939d29-b8dc-4a59-a7b7-37a8675ced9a","Type":"ContainerStarted","Data":"1bdb1c045b9b0a247c6ffd21b6c42c718ccc8e82ceb57e207f6f770ce502e10b"} Oct 09 11:00:29 crc kubenswrapper[4923]: I1009 11:00:29.492247 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:29 crc kubenswrapper[4923]: I1009 11:00:29.492300 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 09 11:00:29 crc kubenswrapper[4923]: I1009 11:00:29.492694 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:30 crc kubenswrapper[4923]: I1009 11:00:30.500981 4923 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.001626 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-37f2-account-create-64wqp" Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.113639 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vjssk\" (UniqueName: \"kubernetes.io/projected/53939d29-b8dc-4a59-a7b7-37a8675ced9a-kube-api-access-vjssk\") pod \"53939d29-b8dc-4a59-a7b7-37a8675ced9a\" (UID: \"53939d29-b8dc-4a59-a7b7-37a8675ced9a\") " Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.131360 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53939d29-b8dc-4a59-a7b7-37a8675ced9a-kube-api-access-vjssk" (OuterVolumeSpecName: "kube-api-access-vjssk") pod "53939d29-b8dc-4a59-a7b7-37a8675ced9a" (UID: "53939d29-b8dc-4a59-a7b7-37a8675ced9a"). InnerVolumeSpecName "kube-api-access-vjssk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.217464 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vjssk\" (UniqueName: \"kubernetes.io/projected/53939d29-b8dc-4a59-a7b7-37a8675ced9a-kube-api-access-vjssk\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.450427 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.511680 4923 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.512824 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-37f2-account-create-64wqp" Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.513089 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-37f2-account-create-64wqp" event={"ID":"53939d29-b8dc-4a59-a7b7-37a8675ced9a","Type":"ContainerDied","Data":"1bdb1c045b9b0a247c6ffd21b6c42c718ccc8e82ceb57e207f6f770ce502e10b"} Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.513129 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bdb1c045b9b0a247c6ffd21b6c42c718ccc8e82ceb57e207f6f770ce502e10b" Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.513204 4923 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 11:00:31 crc kubenswrapper[4923]: I1009 11:00:31.513216 4923 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 11:00:32 crc kubenswrapper[4923]: I1009 11:00:32.153530 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:32 crc kubenswrapper[4923]: I1009 11:00:32.200600 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 09 11:00:32 crc kubenswrapper[4923]: I1009 11:00:32.332203 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.430608 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-z6tpv"] Oct 09 11:00:33 crc kubenswrapper[4923]: E1009 11:00:33.432537 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53939d29-b8dc-4a59-a7b7-37a8675ced9a" containerName="mariadb-account-create" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.432631 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="53939d29-b8dc-4a59-a7b7-37a8675ced9a" containerName="mariadb-account-create" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.432965 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="53939d29-b8dc-4a59-a7b7-37a8675ced9a" containerName="mariadb-account-create" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.433841 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.436725 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.437216 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-9prnt" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.445176 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-z6tpv"] Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.492013 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-combined-ca-bundle\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.492112 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-job-config-data\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.492153 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-config-data\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.492170 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbbtp\" (UniqueName: \"kubernetes.io/projected/e665c786-66db-460a-9bd8-c1ec9ecfb978-kube-api-access-sbbtp\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.596455 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-combined-ca-bundle\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.596928 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-job-config-data\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.597048 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-config-data\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.597088 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbbtp\" (UniqueName: \"kubernetes.io/projected/e665c786-66db-460a-9bd8-c1ec9ecfb978-kube-api-access-sbbtp\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.607015 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-combined-ca-bundle\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.610603 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-config-data\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.612625 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-job-config-data\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.617511 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbbtp\" (UniqueName: \"kubernetes.io/projected/e665c786-66db-460a-9bd8-c1ec9ecfb978-kube-api-access-sbbtp\") pod \"manila-db-sync-z6tpv\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:33 crc kubenswrapper[4923]: I1009 11:00:33.757448 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:34 crc kubenswrapper[4923]: I1009 11:00:34.401028 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-z6tpv"] Oct 09 11:00:34 crc kubenswrapper[4923]: W1009 11:00:34.413659 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode665c786_66db_460a_9bd8_c1ec9ecfb978.slice/crio-10367fa6413e6c172374573368c546a16933243611c127ebd11867b4eebab476 WatchSource:0}: Error finding container 10367fa6413e6c172374573368c546a16933243611c127ebd11867b4eebab476: Status 404 returned error can't find the container with id 10367fa6413e6c172374573368c546a16933243611c127ebd11867b4eebab476 Oct 09 11:00:34 crc kubenswrapper[4923]: I1009 11:00:34.547170 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-z6tpv" event={"ID":"e665c786-66db-460a-9bd8-c1ec9ecfb978","Type":"ContainerStarted","Data":"10367fa6413e6c172374573368c546a16933243611c127ebd11867b4eebab476"} Oct 09 11:00:39 crc kubenswrapper[4923]: I1009 11:00:39.646574 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-z6tpv" event={"ID":"e665c786-66db-460a-9bd8-c1ec9ecfb978","Type":"ContainerStarted","Data":"5ae02b6345a571f707ec2685df5c5510d77e8f74bb5367b11aa75514e34aaf8d"} Oct 09 11:00:39 crc kubenswrapper[4923]: I1009 11:00:39.674143 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-z6tpv" podStartSLOduration=2.209079374 podStartE2EDuration="6.674119613s" podCreationTimestamp="2025-10-09 11:00:33 +0000 UTC" firstStartedPulling="2025-10-09 11:00:34.4179825 +0000 UTC m=+3320.486164256" lastFinishedPulling="2025-10-09 11:00:38.883022739 +0000 UTC m=+3324.951204495" observedRunningTime="2025-10-09 11:00:39.66935111 +0000 UTC m=+3325.737532866" watchObservedRunningTime="2025-10-09 11:00:39.674119613 +0000 UTC m=+3325.742301369" Oct 09 11:00:50 crc kubenswrapper[4923]: I1009 11:00:50.756055 4923 generic.go:334] "Generic (PLEG): container finished" podID="e665c786-66db-460a-9bd8-c1ec9ecfb978" containerID="5ae02b6345a571f707ec2685df5c5510d77e8f74bb5367b11aa75514e34aaf8d" exitCode=0 Oct 09 11:00:50 crc kubenswrapper[4923]: I1009 11:00:50.756168 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-z6tpv" event={"ID":"e665c786-66db-460a-9bd8-c1ec9ecfb978","Type":"ContainerDied","Data":"5ae02b6345a571f707ec2685df5c5510d77e8f74bb5367b11aa75514e34aaf8d"} Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.236451 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.279601 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-combined-ca-bundle\") pod \"e665c786-66db-460a-9bd8-c1ec9ecfb978\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.279921 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-config-data\") pod \"e665c786-66db-460a-9bd8-c1ec9ecfb978\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.280052 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-job-config-data\") pod \"e665c786-66db-460a-9bd8-c1ec9ecfb978\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.280171 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbbtp\" (UniqueName: \"kubernetes.io/projected/e665c786-66db-460a-9bd8-c1ec9ecfb978-kube-api-access-sbbtp\") pod \"e665c786-66db-460a-9bd8-c1ec9ecfb978\" (UID: \"e665c786-66db-460a-9bd8-c1ec9ecfb978\") " Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.287930 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "e665c786-66db-460a-9bd8-c1ec9ecfb978" (UID: "e665c786-66db-460a-9bd8-c1ec9ecfb978"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.288346 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e665c786-66db-460a-9bd8-c1ec9ecfb978-kube-api-access-sbbtp" (OuterVolumeSpecName: "kube-api-access-sbbtp") pod "e665c786-66db-460a-9bd8-c1ec9ecfb978" (UID: "e665c786-66db-460a-9bd8-c1ec9ecfb978"). InnerVolumeSpecName "kube-api-access-sbbtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.292824 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-config-data" (OuterVolumeSpecName: "config-data") pod "e665c786-66db-460a-9bd8-c1ec9ecfb978" (UID: "e665c786-66db-460a-9bd8-c1ec9ecfb978"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.325945 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e665c786-66db-460a-9bd8-c1ec9ecfb978" (UID: "e665c786-66db-460a-9bd8-c1ec9ecfb978"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.383882 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.383929 4923 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-job-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.383950 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbbtp\" (UniqueName: \"kubernetes.io/projected/e665c786-66db-460a-9bd8-c1ec9ecfb978-kube-api-access-sbbtp\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.383963 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e665c786-66db-460a-9bd8-c1ec9ecfb978-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.777135 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-z6tpv" event={"ID":"e665c786-66db-460a-9bd8-c1ec9ecfb978","Type":"ContainerDied","Data":"10367fa6413e6c172374573368c546a16933243611c127ebd11867b4eebab476"} Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.777655 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10367fa6413e6c172374573368c546a16933243611c127ebd11867b4eebab476" Oct 09 11:00:52 crc kubenswrapper[4923]: I1009 11:00:52.777228 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-z6tpv" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.158522 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 09 11:00:53 crc kubenswrapper[4923]: E1009 11:00:53.159276 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e665c786-66db-460a-9bd8-c1ec9ecfb978" containerName="manila-db-sync" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.159299 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="e665c786-66db-460a-9bd8-c1ec9ecfb978" containerName="manila-db-sync" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.159514 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="e665c786-66db-460a-9bd8-c1ec9ecfb978" containerName="manila-db-sync" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.161729 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.167774 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.168034 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.168282 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.168680 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-9prnt" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.195600 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.205940 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8c12316-951b-48ab-a614-7a6ba90b7be8-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.206009 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-scripts\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.206050 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.206150 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55zht\" (UniqueName: \"kubernetes.io/projected/a8c12316-951b-48ab-a614-7a6ba90b7be8-kube-api-access-55zht\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.206292 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.206328 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.217297 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.219145 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.225712 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.252000 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310287 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310376 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310442 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55zht\" (UniqueName: \"kubernetes.io/projected/a8c12316-951b-48ab-a614-7a6ba90b7be8-kube-api-access-55zht\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310482 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310504 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-ceph\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310537 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdqc6\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-kube-api-access-wdqc6\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310573 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310594 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310634 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310656 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310674 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310710 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-scripts\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310731 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8c12316-951b-48ab-a614-7a6ba90b7be8-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.310773 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-scripts\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.317077 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8c12316-951b-48ab-a614-7a6ba90b7be8-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.322636 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.326200 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.326885 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.328500 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-scripts\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.336075 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55zht\" (UniqueName: \"kubernetes.io/projected/a8c12316-951b-48ab-a614-7a6ba90b7be8-kube-api-access-55zht\") pod \"manila-scheduler-0\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.353031 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55d8975557-gm28r"] Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.355824 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.378160 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55d8975557-gm28r"] Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414043 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414126 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-scripts\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414204 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm244\" (UniqueName: \"kubernetes.io/projected/0cccdeaa-06bb-465b-8354-79074c54ada9-kube-api-access-rm244\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414231 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414274 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-ovsdbserver-nb\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414296 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-config\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414312 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-dns-svc\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414363 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414388 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-ceph\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414434 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdqc6\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-kube-api-access-wdqc6\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414475 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-ovsdbserver-sb\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414505 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414535 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.414573 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-openstack-edpm-ipam\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.416038 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.416241 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.420928 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.422509 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.431093 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-scripts\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.431829 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-ceph\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.431861 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.440283 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdqc6\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-kube-api-access-wdqc6\") pod \"manila-share-share1-0\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.480721 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.482818 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.487009 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.487256 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.501582 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.516609 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm244\" (UniqueName: \"kubernetes.io/projected/0cccdeaa-06bb-465b-8354-79074c54ada9-kube-api-access-rm244\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.516738 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-ovsdbserver-nb\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.516797 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-config\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.516822 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-dns-svc\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.518575 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-ovsdbserver-nb\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.518868 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-ovsdbserver-sb\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.519529 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-config\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.517866 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-ovsdbserver-sb\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.519716 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-openstack-edpm-ipam\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.520829 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-openstack-edpm-ipam\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.521243 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0cccdeaa-06bb-465b-8354-79074c54ada9-dns-svc\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.546203 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm244\" (UniqueName: \"kubernetes.io/projected/0cccdeaa-06bb-465b-8354-79074c54ada9-kube-api-access-rm244\") pod \"dnsmasq-dns-55d8975557-gm28r\" (UID: \"0cccdeaa-06bb-465b-8354-79074c54ada9\") " pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.556791 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.610688 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.627115 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.627166 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data-custom\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.627268 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.627286 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-logs\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.627313 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-etc-machine-id\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.627342 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-scripts\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.627391 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s54d\" (UniqueName: \"kubernetes.io/projected/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-kube-api-access-5s54d\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.732476 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.732555 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data-custom\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.732706 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.732740 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-logs\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.732845 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-etc-machine-id\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.732891 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-scripts\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.733199 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s54d\" (UniqueName: \"kubernetes.io/projected/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-kube-api-access-5s54d\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.734590 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-etc-machine-id\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.736523 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-logs\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.743588 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.744292 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data-custom\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.744408 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-scripts\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.750686 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.756360 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s54d\" (UniqueName: \"kubernetes.io/projected/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-kube-api-access-5s54d\") pod \"manila-api-0\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " pod="openstack/manila-api-0" Oct 09 11:00:53 crc kubenswrapper[4923]: I1009 11:00:53.951198 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.133408 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 09 11:00:54 crc kubenswrapper[4923]: W1009 11:00:54.138888 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda8c12316_951b_48ab_a614_7a6ba90b7be8.slice/crio-64d2e18c94afc2ef0dc80dbb0f4bc834dcf39d11f46c9a1397f7468d108c41f4 WatchSource:0}: Error finding container 64d2e18c94afc2ef0dc80dbb0f4bc834dcf39d11f46c9a1397f7468d108c41f4: Status 404 returned error can't find the container with id 64d2e18c94afc2ef0dc80dbb0f4bc834dcf39d11f46c9a1397f7468d108c41f4 Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.285553 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55d8975557-gm28r"] Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.372820 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.578161 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.609995 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.610274 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.828940 4923 generic.go:334] "Generic (PLEG): container finished" podID="0cccdeaa-06bb-465b-8354-79074c54ada9" containerID="0766c250c8f9ece3afa131da58f155800e62eb21e7df1332734657ec76c1b11b" exitCode=0 Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.829071 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d8975557-gm28r" event={"ID":"0cccdeaa-06bb-465b-8354-79074c54ada9","Type":"ContainerDied","Data":"0766c250c8f9ece3afa131da58f155800e62eb21e7df1332734657ec76c1b11b"} Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.829550 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d8975557-gm28r" event={"ID":"0cccdeaa-06bb-465b-8354-79074c54ada9","Type":"ContainerStarted","Data":"b633e5ecacc00534e1b46261b519ac8379e56292cfb79e8e39de6fe82e2fbe65"} Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.835495 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75a2a93f-e98a-496b-a0b3-a3c153a91bcb","Type":"ContainerStarted","Data":"ea5f3cdf4c4cd3625bc408bac8dae7897db947bd088b970920563eb183240a42"} Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.842191 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a8c12316-951b-48ab-a614-7a6ba90b7be8","Type":"ContainerStarted","Data":"64d2e18c94afc2ef0dc80dbb0f4bc834dcf39d11f46c9a1397f7468d108c41f4"} Oct 09 11:00:54 crc kubenswrapper[4923]: I1009 11:00:54.859437 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b8a91455-8e4d-45c3-8796-c4c4cd478f15","Type":"ContainerStarted","Data":"6516d288117a1f53ca55e64852800703c89803a252021b245ffc6229da1396c1"} Oct 09 11:00:55 crc kubenswrapper[4923]: I1009 11:00:55.879336 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55d8975557-gm28r" event={"ID":"0cccdeaa-06bb-465b-8354-79074c54ada9","Type":"ContainerStarted","Data":"a88461bd358309ce73d748b237d439ee50e100e0615db387d3c67734f0984a56"} Oct 09 11:00:55 crc kubenswrapper[4923]: I1009 11:00:55.880086 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:00:55 crc kubenswrapper[4923]: I1009 11:00:55.884738 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75a2a93f-e98a-496b-a0b3-a3c153a91bcb","Type":"ContainerStarted","Data":"5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7"} Oct 09 11:00:55 crc kubenswrapper[4923]: I1009 11:00:55.887451 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a8c12316-951b-48ab-a614-7a6ba90b7be8","Type":"ContainerStarted","Data":"257301b6fe39e5e129dd16f14301dc272d7380c95595a8a7e650a906874007d2"} Oct 09 11:00:55 crc kubenswrapper[4923]: I1009 11:00:55.902491 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55d8975557-gm28r" podStartSLOduration=2.902468225 podStartE2EDuration="2.902468225s" podCreationTimestamp="2025-10-09 11:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 11:00:55.89796271 +0000 UTC m=+3341.966144476" watchObservedRunningTime="2025-10-09 11:00:55.902468225 +0000 UTC m=+3341.970649981" Oct 09 11:00:56 crc kubenswrapper[4923]: I1009 11:00:56.414688 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 09 11:00:56 crc kubenswrapper[4923]: I1009 11:00:56.912786 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75a2a93f-e98a-496b-a0b3-a3c153a91bcb","Type":"ContainerStarted","Data":"8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10"} Oct 09 11:00:56 crc kubenswrapper[4923]: I1009 11:00:56.913255 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 09 11:00:56 crc kubenswrapper[4923]: I1009 11:00:56.928811 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a8c12316-951b-48ab-a614-7a6ba90b7be8","Type":"ContainerStarted","Data":"66bbb4399819f39c0d85ce3ad2fb3ae438d662ea1465ddfac6dbeacbdd7dbcdd"} Oct 09 11:00:56 crc kubenswrapper[4923]: I1009 11:00:56.993010 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.9929648 podStartE2EDuration="3.9929648s" podCreationTimestamp="2025-10-09 11:00:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 11:00:56.952077017 +0000 UTC m=+3343.020258783" watchObservedRunningTime="2025-10-09 11:00:56.9929648 +0000 UTC m=+3343.061146556" Oct 09 11:00:56 crc kubenswrapper[4923]: I1009 11:00:56.997478 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.196074655 podStartE2EDuration="3.997453365s" podCreationTimestamp="2025-10-09 11:00:53 +0000 UTC" firstStartedPulling="2025-10-09 11:00:54.145908182 +0000 UTC m=+3340.214089938" lastFinishedPulling="2025-10-09 11:00:54.947286892 +0000 UTC m=+3341.015468648" observedRunningTime="2025-10-09 11:00:56.990885924 +0000 UTC m=+3343.059067690" watchObservedRunningTime="2025-10-09 11:00:56.997453365 +0000 UTC m=+3343.065635131" Oct 09 11:00:57 crc kubenswrapper[4923]: I1009 11:00:57.965706 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerName="manila-api-log" containerID="cri-o://5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7" gracePeriod=30 Oct 09 11:00:57 crc kubenswrapper[4923]: I1009 11:00:57.966492 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerName="manila-api" containerID="cri-o://8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10" gracePeriod=30 Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.737894 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.801255 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-logs\") pod \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.801443 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-combined-ca-bundle\") pod \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.801475 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-scripts\") pod \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.801512 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data-custom\") pod \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.801574 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data\") pod \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.801614 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s54d\" (UniqueName: \"kubernetes.io/projected/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-kube-api-access-5s54d\") pod \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.801774 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-etc-machine-id\") pod \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\" (UID: \"75a2a93f-e98a-496b-a0b3-a3c153a91bcb\") " Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.802368 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "75a2a93f-e98a-496b-a0b3-a3c153a91bcb" (UID: "75a2a93f-e98a-496b-a0b3-a3c153a91bcb"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.803007 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-logs" (OuterVolumeSpecName: "logs") pod "75a2a93f-e98a-496b-a0b3-a3c153a91bcb" (UID: "75a2a93f-e98a-496b-a0b3-a3c153a91bcb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.813798 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-kube-api-access-5s54d" (OuterVolumeSpecName: "kube-api-access-5s54d") pod "75a2a93f-e98a-496b-a0b3-a3c153a91bcb" (UID: "75a2a93f-e98a-496b-a0b3-a3c153a91bcb"). InnerVolumeSpecName "kube-api-access-5s54d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.815932 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "75a2a93f-e98a-496b-a0b3-a3c153a91bcb" (UID: "75a2a93f-e98a-496b-a0b3-a3c153a91bcb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.819914 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-scripts" (OuterVolumeSpecName: "scripts") pod "75a2a93f-e98a-496b-a0b3-a3c153a91bcb" (UID: "75a2a93f-e98a-496b-a0b3-a3c153a91bcb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.844114 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "75a2a93f-e98a-496b-a0b3-a3c153a91bcb" (UID: "75a2a93f-e98a-496b-a0b3-a3c153a91bcb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.904482 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.904986 4923 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-logs\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.904998 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.905009 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.905017 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.905026 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s54d\" (UniqueName: \"kubernetes.io/projected/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-kube-api-access-5s54d\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.904538 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data" (OuterVolumeSpecName: "config-data") pod "75a2a93f-e98a-496b-a0b3-a3c153a91bcb" (UID: "75a2a93f-e98a-496b-a0b3-a3c153a91bcb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.982151 4923 generic.go:334] "Generic (PLEG): container finished" podID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerID="8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10" exitCode=0 Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.982197 4923 generic.go:334] "Generic (PLEG): container finished" podID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerID="5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7" exitCode=143 Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.982224 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75a2a93f-e98a-496b-a0b3-a3c153a91bcb","Type":"ContainerDied","Data":"8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10"} Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.982264 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75a2a93f-e98a-496b-a0b3-a3c153a91bcb","Type":"ContainerDied","Data":"5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7"} Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.982277 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"75a2a93f-e98a-496b-a0b3-a3c153a91bcb","Type":"ContainerDied","Data":"ea5f3cdf4c4cd3625bc408bac8dae7897db947bd088b970920563eb183240a42"} Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.982283 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 09 11:00:58 crc kubenswrapper[4923]: I1009 11:00:58.982295 4923 scope.go:117] "RemoveContainer" containerID="8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.017558 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75a2a93f-e98a-496b-a0b3-a3c153a91bcb-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.022472 4923 scope.go:117] "RemoveContainer" containerID="5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.026261 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.042043 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.089462 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Oct 09 11:00:59 crc kubenswrapper[4923]: E1009 11:00:59.090059 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerName="manila-api-log" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.090077 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerName="manila-api-log" Oct 09 11:00:59 crc kubenswrapper[4923]: E1009 11:00:59.090118 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerName="manila-api" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.090125 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerName="manila-api" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.090312 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerName="manila-api-log" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.090327 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" containerName="manila-api" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.091483 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.099869 4923 scope.go:117] "RemoveContainer" containerID="8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10" Oct 09 11:00:59 crc kubenswrapper[4923]: E1009 11:00:59.101464 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10\": container with ID starting with 8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10 not found: ID does not exist" containerID="8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.101519 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10"} err="failed to get container status \"8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10\": rpc error: code = NotFound desc = could not find container \"8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10\": container with ID starting with 8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10 not found: ID does not exist" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.101554 4923 scope.go:117] "RemoveContainer" containerID="5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7" Oct 09 11:00:59 crc kubenswrapper[4923]: E1009 11:00:59.102411 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7\": container with ID starting with 5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7 not found: ID does not exist" containerID="5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.102442 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7"} err="failed to get container status \"5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7\": rpc error: code = NotFound desc = could not find container \"5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7\": container with ID starting with 5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7 not found: ID does not exist" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.102467 4923 scope.go:117] "RemoveContainer" containerID="8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.102990 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.103205 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10"} err="failed to get container status \"8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10\": rpc error: code = NotFound desc = could not find container \"8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10\": container with ID starting with 8465c7ad953679784cf650ac8d552d9e6dd3321be096a88a84d8323a7663ff10 not found: ID does not exist" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.103240 4923 scope.go:117] "RemoveContainer" containerID="5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.102996 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.103992 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7"} err="failed to get container status \"5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7\": rpc error: code = NotFound desc = could not find container \"5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7\": container with ID starting with 5e49f7273c60217cde425e25b13bcc907350a5c418acd8e2039ff055cadb21d7 not found: ID does not exist" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.104888 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.110014 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.221895 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-internal-tls-certs\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.221972 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0afc3325-d376-42d0-a3b2-650f4f707d56-logs\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.222023 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-public-tls-certs\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.222055 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dswmb\" (UniqueName: \"kubernetes.io/projected/0afc3325-d376-42d0-a3b2-650f4f707d56-kube-api-access-dswmb\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.222075 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-config-data-custom\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.222096 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0afc3325-d376-42d0-a3b2-650f4f707d56-etc-machine-id\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.222400 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-config-data\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.222640 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-scripts\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.222895 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.324968 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.325360 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-internal-tls-certs\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.325512 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0afc3325-d376-42d0-a3b2-650f4f707d56-logs\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.325618 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-public-tls-certs\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.325733 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dswmb\" (UniqueName: \"kubernetes.io/projected/0afc3325-d376-42d0-a3b2-650f4f707d56-kube-api-access-dswmb\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.326332 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-config-data-custom\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.326443 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0afc3325-d376-42d0-a3b2-650f4f707d56-etc-machine-id\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.326609 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-config-data\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.326735 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-scripts\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.330031 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-internal-tls-certs\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.326018 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0afc3325-d376-42d0-a3b2-650f4f707d56-logs\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.330177 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0afc3325-d376-42d0-a3b2-650f4f707d56-etc-machine-id\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.333591 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-config-data-custom\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.334169 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-scripts\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.334881 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.344406 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.344738 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="ceilometer-central-agent" containerID="cri-o://0a26caf24f63cdb362fad2b04c5f6032e42d26a46470aa8536c2d700a687d53d" gracePeriod=30 Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.345249 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="proxy-httpd" containerID="cri-o://9befc1b806d4460d9d89e18548c3d0a5a59c230dc9eabfb969a1367cb2ee88ab" gracePeriod=30 Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.345305 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="sg-core" containerID="cri-o://599134c26d70dc49eb13fc37f78e169840223afab406f931c906a7ea090e5b92" gracePeriod=30 Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.345342 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="ceilometer-notification-agent" containerID="cri-o://96b37b65285ebc87dfd6573dd7e0528c2ad8cb41af2adb2576cefa9387f7bf55" gracePeriod=30 Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.346586 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-public-tls-certs\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.349976 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0afc3325-d376-42d0-a3b2-650f4f707d56-config-data\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.357923 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dswmb\" (UniqueName: \"kubernetes.io/projected/0afc3325-d376-42d0-a3b2-650f4f707d56-kube-api-access-dswmb\") pod \"manila-api-0\" (UID: \"0afc3325-d376-42d0-a3b2-650f4f707d56\") " pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.462868 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.999474 4923 generic.go:334] "Generic (PLEG): container finished" podID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerID="9befc1b806d4460d9d89e18548c3d0a5a59c230dc9eabfb969a1367cb2ee88ab" exitCode=0 Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.999512 4923 generic.go:334] "Generic (PLEG): container finished" podID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerID="599134c26d70dc49eb13fc37f78e169840223afab406f931c906a7ea090e5b92" exitCode=2 Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.999519 4923 generic.go:334] "Generic (PLEG): container finished" podID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerID="0a26caf24f63cdb362fad2b04c5f6032e42d26a46470aa8536c2d700a687d53d" exitCode=0 Oct 09 11:00:59 crc kubenswrapper[4923]: I1009 11:00:59.999539 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerDied","Data":"9befc1b806d4460d9d89e18548c3d0a5a59c230dc9eabfb969a1367cb2ee88ab"} Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:00:59.999566 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerDied","Data":"599134c26d70dc49eb13fc37f78e169840223afab406f931c906a7ea090e5b92"} Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:00:59.999576 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerDied","Data":"0a26caf24f63cdb362fad2b04c5f6032e42d26a46470aa8536c2d700a687d53d"} Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.149160 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29333461-8cpkk"] Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.151051 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.183434 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29333461-8cpkk"] Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.246689 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-fernet-keys\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.246857 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-combined-ca-bundle\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.246885 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-config-data\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.247113 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22z5z\" (UniqueName: \"kubernetes.io/projected/2917616d-7104-4c9e-9fa2-2590c69866b2-kube-api-access-22z5z\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.349500 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-combined-ca-bundle\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.349565 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-config-data\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.349638 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22z5z\" (UniqueName: \"kubernetes.io/projected/2917616d-7104-4c9e-9fa2-2590c69866b2-kube-api-access-22z5z\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.349746 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-fernet-keys\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.358167 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-combined-ca-bundle\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.358464 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-fernet-keys\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.368120 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-config-data\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.371586 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22z5z\" (UniqueName: \"kubernetes.io/projected/2917616d-7104-4c9e-9fa2-2590c69866b2-kube-api-access-22z5z\") pod \"keystone-cron-29333461-8cpkk\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.482416 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:00 crc kubenswrapper[4923]: I1009 11:01:00.615356 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75a2a93f-e98a-496b-a0b3-a3c153a91bcb" path="/var/lib/kubelet/pods/75a2a93f-e98a-496b-a0b3-a3c153a91bcb/volumes" Oct 09 11:01:01 crc kubenswrapper[4923]: I1009 11:01:01.128133 4923 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.192:3000/\": dial tcp 10.217.0.192:3000: connect: connection refused" Oct 09 11:01:02 crc kubenswrapper[4923]: I1009 11:01:02.028048 4923 generic.go:334] "Generic (PLEG): container finished" podID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerID="96b37b65285ebc87dfd6573dd7e0528c2ad8cb41af2adb2576cefa9387f7bf55" exitCode=0 Oct 09 11:01:02 crc kubenswrapper[4923]: I1009 11:01:02.028104 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerDied","Data":"96b37b65285ebc87dfd6573dd7e0528c2ad8cb41af2adb2576cefa9387f7bf55"} Oct 09 11:01:02 crc kubenswrapper[4923]: I1009 11:01:02.960392 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.046481 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvpjs\" (UniqueName: \"kubernetes.io/projected/88b5043d-77b0-40c5-b0d0-c9283f084046-kube-api-access-bvpjs\") pod \"88b5043d-77b0-40c5-b0d0-c9283f084046\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.046562 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-scripts\") pod \"88b5043d-77b0-40c5-b0d0-c9283f084046\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.046647 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-log-httpd\") pod \"88b5043d-77b0-40c5-b0d0-c9283f084046\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.046669 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-run-httpd\") pod \"88b5043d-77b0-40c5-b0d0-c9283f084046\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.046710 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-combined-ca-bundle\") pod \"88b5043d-77b0-40c5-b0d0-c9283f084046\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.046814 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-config-data\") pod \"88b5043d-77b0-40c5-b0d0-c9283f084046\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.046972 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-sg-core-conf-yaml\") pod \"88b5043d-77b0-40c5-b0d0-c9283f084046\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.047007 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-ceilometer-tls-certs\") pod \"88b5043d-77b0-40c5-b0d0-c9283f084046\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.050355 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "88b5043d-77b0-40c5-b0d0-c9283f084046" (UID: "88b5043d-77b0-40c5-b0d0-c9283f084046"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.050842 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "88b5043d-77b0-40c5-b0d0-c9283f084046" (UID: "88b5043d-77b0-40c5-b0d0-c9283f084046"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.062990 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-scripts" (OuterVolumeSpecName: "scripts") pod "88b5043d-77b0-40c5-b0d0-c9283f084046" (UID: "88b5043d-77b0-40c5-b0d0-c9283f084046"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.068137 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88b5043d-77b0-40c5-b0d0-c9283f084046-kube-api-access-bvpjs" (OuterVolumeSpecName: "kube-api-access-bvpjs") pod "88b5043d-77b0-40c5-b0d0-c9283f084046" (UID: "88b5043d-77b0-40c5-b0d0-c9283f084046"). InnerVolumeSpecName "kube-api-access-bvpjs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.081633 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"88b5043d-77b0-40c5-b0d0-c9283f084046","Type":"ContainerDied","Data":"774aae2b684e8d8c4c268bd7ce85b1a317b7b4a4f0dd3d5b80bf14bc2c3851ca"} Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.081702 4923 scope.go:117] "RemoveContainer" containerID="9befc1b806d4460d9d89e18548c3d0a5a59c230dc9eabfb969a1367cb2ee88ab" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.081870 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.110263 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "88b5043d-77b0-40c5-b0d0-c9283f084046" (UID: "88b5043d-77b0-40c5-b0d0-c9283f084046"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.151222 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvpjs\" (UniqueName: \"kubernetes.io/projected/88b5043d-77b0-40c5-b0d0-c9283f084046-kube-api-access-bvpjs\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.151285 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.151304 4923 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.151320 4923 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/88b5043d-77b0-40c5-b0d0-c9283f084046-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.151336 4923 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.165405 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "88b5043d-77b0-40c5-b0d0-c9283f084046" (UID: "88b5043d-77b0-40c5-b0d0-c9283f084046"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.174201 4923 scope.go:117] "RemoveContainer" containerID="599134c26d70dc49eb13fc37f78e169840223afab406f931c906a7ea090e5b92" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.214407 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29333461-8cpkk"] Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.225874 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-config-data" (OuterVolumeSpecName: "config-data") pod "88b5043d-77b0-40c5-b0d0-c9283f084046" (UID: "88b5043d-77b0-40c5-b0d0-c9283f084046"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.240968 4923 scope.go:117] "RemoveContainer" containerID="96b37b65285ebc87dfd6573dd7e0528c2ad8cb41af2adb2576cefa9387f7bf55" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.251935 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88b5043d-77b0-40c5-b0d0-c9283f084046" (UID: "88b5043d-77b0-40c5-b0d0-c9283f084046"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.253021 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-combined-ca-bundle\") pod \"88b5043d-77b0-40c5-b0d0-c9283f084046\" (UID: \"88b5043d-77b0-40c5-b0d0-c9283f084046\") " Oct 09 11:01:03 crc kubenswrapper[4923]: W1009 11:01:03.254466 4923 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/88b5043d-77b0-40c5-b0d0-c9283f084046/volumes/kubernetes.io~secret/combined-ca-bundle Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.254509 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88b5043d-77b0-40c5-b0d0-c9283f084046" (UID: "88b5043d-77b0-40c5-b0d0-c9283f084046"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.254670 4923 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.254701 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.254716 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88b5043d-77b0-40c5-b0d0-c9283f084046-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.299697 4923 scope.go:117] "RemoveContainer" containerID="0a26caf24f63cdb362fad2b04c5f6032e42d26a46470aa8536c2d700a687d53d" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.306867 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Oct 09 11:01:03 crc kubenswrapper[4923]: W1009 11:01:03.307985 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0afc3325_d376_42d0_a3b2_650f4f707d56.slice/crio-5b2db688696f0ad32ff2398611840a684738b083ffb04c1bac4034d5eb46e8dc WatchSource:0}: Error finding container 5b2db688696f0ad32ff2398611840a684738b083ffb04c1bac4034d5eb46e8dc: Status 404 returned error can't find the container with id 5b2db688696f0ad32ff2398611840a684738b083ffb04c1bac4034d5eb46e8dc Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.432633 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.451897 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.467579 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:03 crc kubenswrapper[4923]: E1009 11:01:03.468119 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="proxy-httpd" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.468142 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="proxy-httpd" Oct 09 11:01:03 crc kubenswrapper[4923]: E1009 11:01:03.468169 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="ceilometer-central-agent" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.468176 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="ceilometer-central-agent" Oct 09 11:01:03 crc kubenswrapper[4923]: E1009 11:01:03.468199 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="sg-core" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.468206 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="sg-core" Oct 09 11:01:03 crc kubenswrapper[4923]: E1009 11:01:03.468221 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="ceilometer-notification-agent" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.468227 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="ceilometer-notification-agent" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.468426 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="ceilometer-central-agent" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.468447 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="ceilometer-notification-agent" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.468459 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="proxy-httpd" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.468478 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" containerName="sg-core" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.470910 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.473346 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.474884 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.475420 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.482808 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.500303 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.561719 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-config-data\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.561812 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.561865 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.561916 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5jd2\" (UniqueName: \"kubernetes.io/projected/277b1f0f-f601-4de4-92c8-8b44b9e1da33-kube-api-access-m5jd2\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.562007 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-scripts\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.562034 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-log-httpd\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.562060 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.562087 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-run-httpd\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.614168 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55d8975557-gm28r" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.665057 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-scripts\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.665127 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-log-httpd\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.665166 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.665205 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-run-httpd\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.665259 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-config-data\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.665288 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.665328 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.665377 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5jd2\" (UniqueName: \"kubernetes.io/projected/277b1f0f-f601-4de4-92c8-8b44b9e1da33-kube-api-access-m5jd2\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.667395 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-run-httpd\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.668523 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-log-httpd\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.672961 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.678242 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.683944 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-config-data\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.684235 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-scripts\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.686004 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.708258 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f5d87575-vkntz"] Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.708564 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" podUID="1fd11e59-1ac0-4158-b2f9-2286825f3979" containerName="dnsmasq-dns" containerID="cri-o://48d3c93b11239e1079e647ef75ae7350edd86932e1c1e41d20c986dc67050ccb" gracePeriod=10 Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.713605 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5jd2\" (UniqueName: \"kubernetes.io/projected/277b1f0f-f601-4de4-92c8-8b44b9e1da33-kube-api-access-m5jd2\") pod \"ceilometer-0\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " pod="openstack/ceilometer-0" Oct 09 11:01:03 crc kubenswrapper[4923]: I1009 11:01:03.812424 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.133318 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0afc3325-d376-42d0-a3b2-650f4f707d56","Type":"ContainerStarted","Data":"54545268119e032fb8e68ae66ca1f5b0f40bf4e7fc9fb3acd8aff9a09d777bfc"} Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.133786 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0afc3325-d376-42d0-a3b2-650f4f707d56","Type":"ContainerStarted","Data":"5b2db688696f0ad32ff2398611840a684738b083ffb04c1bac4034d5eb46e8dc"} Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.147395 4923 generic.go:334] "Generic (PLEG): container finished" podID="1fd11e59-1ac0-4158-b2f9-2286825f3979" containerID="48d3c93b11239e1079e647ef75ae7350edd86932e1c1e41d20c986dc67050ccb" exitCode=0 Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.147841 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" event={"ID":"1fd11e59-1ac0-4158-b2f9-2286825f3979","Type":"ContainerDied","Data":"48d3c93b11239e1079e647ef75ae7350edd86932e1c1e41d20c986dc67050ccb"} Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.161549 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29333461-8cpkk" event={"ID":"2917616d-7104-4c9e-9fa2-2590c69866b2","Type":"ContainerStarted","Data":"ec8dab2f2ded905735026e6d451cfe4994651e3d60b6e2c7a432c5e7d2449c19"} Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.161612 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29333461-8cpkk" event={"ID":"2917616d-7104-4c9e-9fa2-2590c69866b2","Type":"ContainerStarted","Data":"ccac66527777a07698ce82aa79dec2c049adab7f813660352e199811703fb504"} Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.169169 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b8a91455-8e4d-45c3-8796-c4c4cd478f15","Type":"ContainerStarted","Data":"642eca6c1f47e753b16976e49881bcb864786e172f40c4a0280d0e7464ddc1dc"} Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.187514 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29333461-8cpkk" podStartSLOduration=4.187484118 podStartE2EDuration="4.187484118s" podCreationTimestamp="2025-10-09 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 11:01:04.182505931 +0000 UTC m=+3350.250687687" watchObservedRunningTime="2025-10-09 11:01:04.187484118 +0000 UTC m=+3350.255665874" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.377531 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.493788 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-sb\") pod \"1fd11e59-1ac0-4158-b2f9-2286825f3979\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.493915 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc5v9\" (UniqueName: \"kubernetes.io/projected/1fd11e59-1ac0-4158-b2f9-2286825f3979-kube-api-access-pc5v9\") pod \"1fd11e59-1ac0-4158-b2f9-2286825f3979\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.494001 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-nb\") pod \"1fd11e59-1ac0-4158-b2f9-2286825f3979\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.494082 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-openstack-edpm-ipam\") pod \"1fd11e59-1ac0-4158-b2f9-2286825f3979\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.494162 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-config\") pod \"1fd11e59-1ac0-4158-b2f9-2286825f3979\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.494236 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-dns-svc\") pod \"1fd11e59-1ac0-4158-b2f9-2286825f3979\" (UID: \"1fd11e59-1ac0-4158-b2f9-2286825f3979\") " Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.513274 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fd11e59-1ac0-4158-b2f9-2286825f3979-kube-api-access-pc5v9" (OuterVolumeSpecName: "kube-api-access-pc5v9") pod "1fd11e59-1ac0-4158-b2f9-2286825f3979" (UID: "1fd11e59-1ac0-4158-b2f9-2286825f3979"). InnerVolumeSpecName "kube-api-access-pc5v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.597594 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.626437 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc5v9\" (UniqueName: \"kubernetes.io/projected/1fd11e59-1ac0-4158-b2f9-2286825f3979-kube-api-access-pc5v9\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.671392 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1fd11e59-1ac0-4158-b2f9-2286825f3979" (UID: "1fd11e59-1ac0-4158-b2f9-2286825f3979"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.720484 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-config" (OuterVolumeSpecName: "config") pod "1fd11e59-1ac0-4158-b2f9-2286825f3979" (UID: "1fd11e59-1ac0-4158-b2f9-2286825f3979"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.731855 4923 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-config\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.731897 4923 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.752157 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1fd11e59-1ac0-4158-b2f9-2286825f3979" (UID: "1fd11e59-1ac0-4158-b2f9-2286825f3979"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.766241 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88b5043d-77b0-40c5-b0d0-c9283f084046" path="/var/lib/kubelet/pods/88b5043d-77b0-40c5-b0d0-c9283f084046/volumes" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.803006 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "1fd11e59-1ac0-4158-b2f9-2286825f3979" (UID: "1fd11e59-1ac0-4158-b2f9-2286825f3979"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.834040 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.834087 4923 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.847428 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1fd11e59-1ac0-4158-b2f9-2286825f3979" (UID: "1fd11e59-1ac0-4158-b2f9-2286825f3979"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 11:01:04 crc kubenswrapper[4923]: I1009 11:01:04.936623 4923 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1fd11e59-1ac0-4158-b2f9-2286825f3979-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.204413 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"0afc3325-d376-42d0-a3b2-650f4f707d56","Type":"ContainerStarted","Data":"e9430376373f96685ab9eb04be45c9617f485c4c85cbbe2a737c07779bcf783c"} Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.204933 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.213265 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.213205 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f5d87575-vkntz" event={"ID":"1fd11e59-1ac0-4158-b2f9-2286825f3979","Type":"ContainerDied","Data":"cac03d949bcbffb8cb9ad15776309518356bed3840e25f4df126d5fcc958b1f9"} Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.214096 4923 scope.go:117] "RemoveContainer" containerID="48d3c93b11239e1079e647ef75ae7350edd86932e1c1e41d20c986dc67050ccb" Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.220464 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerStarted","Data":"7a6b89e783669a1282687c5165ac06b2ba2e1924b4f939be13f0cadb54597c62"} Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.236725 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=6.236699529 podStartE2EDuration="6.236699529s" podCreationTimestamp="2025-10-09 11:00:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 11:01:05.221305053 +0000 UTC m=+3351.289486809" watchObservedRunningTime="2025-10-09 11:01:05.236699529 +0000 UTC m=+3351.304881285" Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.242962 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b8a91455-8e4d-45c3-8796-c4c4cd478f15","Type":"ContainerStarted","Data":"87189fc15435f69db9f038a33a12f39020ef7741907395ab9bacdec10d435786"} Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.256470 4923 scope.go:117] "RemoveContainer" containerID="4d68ac044f82edb0a9fe5ff23743b1a235111e05923378e0215b34c12c1d6fdc" Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.271822 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.984677092 podStartE2EDuration="12.271795753s" podCreationTimestamp="2025-10-09 11:00:53 +0000 UTC" firstStartedPulling="2025-10-09 11:00:54.394352821 +0000 UTC m=+3340.462534577" lastFinishedPulling="2025-10-09 11:01:02.681471482 +0000 UTC m=+3348.749653238" observedRunningTime="2025-10-09 11:01:05.267884954 +0000 UTC m=+3351.336066720" watchObservedRunningTime="2025-10-09 11:01:05.271795753 +0000 UTC m=+3351.339977509" Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.313191 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f5d87575-vkntz"] Oct 09 11:01:05 crc kubenswrapper[4923]: I1009 11:01:05.338825 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f5d87575-vkntz"] Oct 09 11:01:06 crc kubenswrapper[4923]: I1009 11:01:06.254357 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerStarted","Data":"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec"} Oct 09 11:01:06 crc kubenswrapper[4923]: I1009 11:01:06.619371 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fd11e59-1ac0-4158-b2f9-2286825f3979" path="/var/lib/kubelet/pods/1fd11e59-1ac0-4158-b2f9-2286825f3979/volumes" Oct 09 11:01:07 crc kubenswrapper[4923]: I1009 11:01:07.271918 4923 generic.go:334] "Generic (PLEG): container finished" podID="2917616d-7104-4c9e-9fa2-2590c69866b2" containerID="ec8dab2f2ded905735026e6d451cfe4994651e3d60b6e2c7a432c5e7d2449c19" exitCode=0 Oct 09 11:01:07 crc kubenswrapper[4923]: I1009 11:01:07.271995 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29333461-8cpkk" event={"ID":"2917616d-7104-4c9e-9fa2-2590c69866b2","Type":"ContainerDied","Data":"ec8dab2f2ded905735026e6d451cfe4994651e3d60b6e2c7a432c5e7d2449c19"} Oct 09 11:01:07 crc kubenswrapper[4923]: I1009 11:01:07.276811 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerStarted","Data":"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d"} Oct 09 11:01:07 crc kubenswrapper[4923]: I1009 11:01:07.517455 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.291566 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerStarted","Data":"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb"} Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.689847 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.742503 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-fernet-keys\") pod \"2917616d-7104-4c9e-9fa2-2590c69866b2\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.743032 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-combined-ca-bundle\") pod \"2917616d-7104-4c9e-9fa2-2590c69866b2\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.743429 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22z5z\" (UniqueName: \"kubernetes.io/projected/2917616d-7104-4c9e-9fa2-2590c69866b2-kube-api-access-22z5z\") pod \"2917616d-7104-4c9e-9fa2-2590c69866b2\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.743557 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-config-data\") pod \"2917616d-7104-4c9e-9fa2-2590c69866b2\" (UID: \"2917616d-7104-4c9e-9fa2-2590c69866b2\") " Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.814238 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2917616d-7104-4c9e-9fa2-2590c69866b2" (UID: "2917616d-7104-4c9e-9fa2-2590c69866b2"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.821735 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2917616d-7104-4c9e-9fa2-2590c69866b2" (UID: "2917616d-7104-4c9e-9fa2-2590c69866b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.829827 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2917616d-7104-4c9e-9fa2-2590c69866b2-kube-api-access-22z5z" (OuterVolumeSpecName: "kube-api-access-22z5z") pod "2917616d-7104-4c9e-9fa2-2590c69866b2" (UID: "2917616d-7104-4c9e-9fa2-2590c69866b2"). InnerVolumeSpecName "kube-api-access-22z5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.847579 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22z5z\" (UniqueName: \"kubernetes.io/projected/2917616d-7104-4c9e-9fa2-2590c69866b2-kube-api-access-22z5z\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.847620 4923 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.847636 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.853774 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-config-data" (OuterVolumeSpecName: "config-data") pod "2917616d-7104-4c9e-9fa2-2590c69866b2" (UID: "2917616d-7104-4c9e-9fa2-2590c69866b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:08 crc kubenswrapper[4923]: I1009 11:01:08.951823 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2917616d-7104-4c9e-9fa2-2590c69866b2-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:09 crc kubenswrapper[4923]: I1009 11:01:09.308955 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29333461-8cpkk" event={"ID":"2917616d-7104-4c9e-9fa2-2590c69866b2","Type":"ContainerDied","Data":"ccac66527777a07698ce82aa79dec2c049adab7f813660352e199811703fb504"} Oct 09 11:01:09 crc kubenswrapper[4923]: I1009 11:01:09.309004 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccac66527777a07698ce82aa79dec2c049adab7f813660352e199811703fb504" Oct 09 11:01:09 crc kubenswrapper[4923]: I1009 11:01:09.309006 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29333461-8cpkk" Oct 09 11:01:10 crc kubenswrapper[4923]: I1009 11:01:10.339895 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerStarted","Data":"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d"} Oct 09 11:01:10 crc kubenswrapper[4923]: I1009 11:01:10.340586 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 11:01:10 crc kubenswrapper[4923]: I1009 11:01:10.340395 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="sg-core" containerID="cri-o://554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb" gracePeriod=30 Oct 09 11:01:10 crc kubenswrapper[4923]: I1009 11:01:10.340161 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="ceilometer-central-agent" containerID="cri-o://dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec" gracePeriod=30 Oct 09 11:01:10 crc kubenswrapper[4923]: I1009 11:01:10.340468 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="ceilometer-notification-agent" containerID="cri-o://e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d" gracePeriod=30 Oct 09 11:01:10 crc kubenswrapper[4923]: I1009 11:01:10.340444 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="proxy-httpd" containerID="cri-o://28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d" gracePeriod=30 Oct 09 11:01:10 crc kubenswrapper[4923]: I1009 11:01:10.373546 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.553518872 podStartE2EDuration="7.373520534s" podCreationTimestamp="2025-10-09 11:01:03 +0000 UTC" firstStartedPulling="2025-10-09 11:01:04.733045074 +0000 UTC m=+3350.801226830" lastFinishedPulling="2025-10-09 11:01:09.553046736 +0000 UTC m=+3355.621228492" observedRunningTime="2025-10-09 11:01:10.367641451 +0000 UTC m=+3356.435823207" watchObservedRunningTime="2025-10-09 11:01:10.373520534 +0000 UTC m=+3356.441702290" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.149312 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.208164 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5jd2\" (UniqueName: \"kubernetes.io/projected/277b1f0f-f601-4de4-92c8-8b44b9e1da33-kube-api-access-m5jd2\") pod \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.208240 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-scripts\") pod \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.208309 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-sg-core-conf-yaml\") pod \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.208370 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-combined-ca-bundle\") pod \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.208393 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-config-data\") pod \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.208418 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-log-httpd\") pod \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.208456 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-run-httpd\") pod \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.208500 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-ceilometer-tls-certs\") pod \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\" (UID: \"277b1f0f-f601-4de4-92c8-8b44b9e1da33\") " Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.210071 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "277b1f0f-f601-4de4-92c8-8b44b9e1da33" (UID: "277b1f0f-f601-4de4-92c8-8b44b9e1da33"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.214318 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "277b1f0f-f601-4de4-92c8-8b44b9e1da33" (UID: "277b1f0f-f601-4de4-92c8-8b44b9e1da33"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.223667 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/277b1f0f-f601-4de4-92c8-8b44b9e1da33-kube-api-access-m5jd2" (OuterVolumeSpecName: "kube-api-access-m5jd2") pod "277b1f0f-f601-4de4-92c8-8b44b9e1da33" (UID: "277b1f0f-f601-4de4-92c8-8b44b9e1da33"). InnerVolumeSpecName "kube-api-access-m5jd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.235127 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-scripts" (OuterVolumeSpecName: "scripts") pod "277b1f0f-f601-4de4-92c8-8b44b9e1da33" (UID: "277b1f0f-f601-4de4-92c8-8b44b9e1da33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.275259 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "277b1f0f-f601-4de4-92c8-8b44b9e1da33" (UID: "277b1f0f-f601-4de4-92c8-8b44b9e1da33"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.284714 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "277b1f0f-f601-4de4-92c8-8b44b9e1da33" (UID: "277b1f0f-f601-4de4-92c8-8b44b9e1da33"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.311498 4923 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.311548 4923 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.311567 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5jd2\" (UniqueName: \"kubernetes.io/projected/277b1f0f-f601-4de4-92c8-8b44b9e1da33-kube-api-access-m5jd2\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.311579 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.311592 4923 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.311603 4923 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/277b1f0f-f601-4de4-92c8-8b44b9e1da33-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.321830 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "277b1f0f-f601-4de4-92c8-8b44b9e1da33" (UID: "277b1f0f-f601-4de4-92c8-8b44b9e1da33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.329729 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-config-data" (OuterVolumeSpecName: "config-data") pod "277b1f0f-f601-4de4-92c8-8b44b9e1da33" (UID: "277b1f0f-f601-4de4-92c8-8b44b9e1da33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351470 4923 generic.go:334] "Generic (PLEG): container finished" podID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerID="28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d" exitCode=0 Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351513 4923 generic.go:334] "Generic (PLEG): container finished" podID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerID="554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb" exitCode=2 Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351522 4923 generic.go:334] "Generic (PLEG): container finished" podID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerID="e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d" exitCode=0 Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351531 4923 generic.go:334] "Generic (PLEG): container finished" podID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerID="dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec" exitCode=0 Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351531 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerDied","Data":"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d"} Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351605 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerDied","Data":"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb"} Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351621 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerDied","Data":"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d"} Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351633 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerDied","Data":"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec"} Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351645 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"277b1f0f-f601-4de4-92c8-8b44b9e1da33","Type":"ContainerDied","Data":"7a6b89e783669a1282687c5165ac06b2ba2e1924b4f939be13f0cadb54597c62"} Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.351666 4923 scope.go:117] "RemoveContainer" containerID="28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.352811 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.377776 4923 scope.go:117] "RemoveContainer" containerID="554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.403012 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.413778 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.413828 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/277b1f0f-f601-4de4-92c8-8b44b9e1da33-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.421313 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.429259 4923 scope.go:117] "RemoveContainer" containerID="e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.432920 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.433529 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd11e59-1ac0-4158-b2f9-2286825f3979" containerName="init" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433547 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd11e59-1ac0-4158-b2f9-2286825f3979" containerName="init" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.433566 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="sg-core" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433571 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="sg-core" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.433594 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="proxy-httpd" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433600 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="proxy-httpd" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.433620 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2917616d-7104-4c9e-9fa2-2590c69866b2" containerName="keystone-cron" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433626 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="2917616d-7104-4c9e-9fa2-2590c69866b2" containerName="keystone-cron" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.433640 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="ceilometer-central-agent" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433647 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="ceilometer-central-agent" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.433669 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fd11e59-1ac0-4158-b2f9-2286825f3979" containerName="dnsmasq-dns" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433675 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fd11e59-1ac0-4158-b2f9-2286825f3979" containerName="dnsmasq-dns" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.433683 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="ceilometer-notification-agent" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433689 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="ceilometer-notification-agent" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433913 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="ceilometer-central-agent" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433938 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="ceilometer-notification-agent" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433948 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="proxy-httpd" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433959 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="2917616d-7104-4c9e-9fa2-2590c69866b2" containerName="keystone-cron" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433967 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fd11e59-1ac0-4158-b2f9-2286825f3979" containerName="dnsmasq-dns" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.433974 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" containerName="sg-core" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.436050 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.438454 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.438832 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.438976 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.448848 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.465253 4923 scope.go:117] "RemoveContainer" containerID="dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.487712 4923 scope.go:117] "RemoveContainer" containerID="28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.488672 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d\": container with ID starting with 28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d not found: ID does not exist" containerID="28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.488739 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d"} err="failed to get container status \"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d\": rpc error: code = NotFound desc = could not find container \"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d\": container with ID starting with 28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.488793 4923 scope.go:117] "RemoveContainer" containerID="554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.489166 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb\": container with ID starting with 554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb not found: ID does not exist" containerID="554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.489206 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb"} err="failed to get container status \"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb\": rpc error: code = NotFound desc = could not find container \"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb\": container with ID starting with 554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.489225 4923 scope.go:117] "RemoveContainer" containerID="e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.489446 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d\": container with ID starting with e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d not found: ID does not exist" containerID="e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.489464 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d"} err="failed to get container status \"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d\": rpc error: code = NotFound desc = could not find container \"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d\": container with ID starting with e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.489499 4923 scope.go:117] "RemoveContainer" containerID="dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec" Oct 09 11:01:11 crc kubenswrapper[4923]: E1009 11:01:11.489796 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec\": container with ID starting with dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec not found: ID does not exist" containerID="dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.489837 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec"} err="failed to get container status \"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec\": rpc error: code = NotFound desc = could not find container \"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec\": container with ID starting with dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.489850 4923 scope.go:117] "RemoveContainer" containerID="28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.490133 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d"} err="failed to get container status \"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d\": rpc error: code = NotFound desc = could not find container \"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d\": container with ID starting with 28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.490172 4923 scope.go:117] "RemoveContainer" containerID="554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.490357 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb"} err="failed to get container status \"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb\": rpc error: code = NotFound desc = could not find container \"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb\": container with ID starting with 554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.490374 4923 scope.go:117] "RemoveContainer" containerID="e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.490662 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d"} err="failed to get container status \"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d\": rpc error: code = NotFound desc = could not find container \"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d\": container with ID starting with e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.490720 4923 scope.go:117] "RemoveContainer" containerID="dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.491018 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec"} err="failed to get container status \"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec\": rpc error: code = NotFound desc = could not find container \"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec\": container with ID starting with dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.491039 4923 scope.go:117] "RemoveContainer" containerID="28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.491216 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d"} err="failed to get container status \"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d\": rpc error: code = NotFound desc = could not find container \"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d\": container with ID starting with 28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.491233 4923 scope.go:117] "RemoveContainer" containerID="554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.491988 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb"} err="failed to get container status \"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb\": rpc error: code = NotFound desc = could not find container \"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb\": container with ID starting with 554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.492016 4923 scope.go:117] "RemoveContainer" containerID="e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.492366 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d"} err="failed to get container status \"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d\": rpc error: code = NotFound desc = could not find container \"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d\": container with ID starting with e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.492398 4923 scope.go:117] "RemoveContainer" containerID="dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.492739 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec"} err="failed to get container status \"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec\": rpc error: code = NotFound desc = could not find container \"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec\": container with ID starting with dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.492776 4923 scope.go:117] "RemoveContainer" containerID="28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.493173 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d"} err="failed to get container status \"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d\": rpc error: code = NotFound desc = could not find container \"28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d\": container with ID starting with 28157a3603c3f2e205df0e98e7cefdc18e714d119e37578469f8405fd831517d not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.493201 4923 scope.go:117] "RemoveContainer" containerID="554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.493596 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb"} err="failed to get container status \"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb\": rpc error: code = NotFound desc = could not find container \"554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb\": container with ID starting with 554ce4b608ce67248928ff1a7ffc302c7a19a7c6f9f2f0d8e5c322a203d721cb not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.493628 4923 scope.go:117] "RemoveContainer" containerID="e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.494043 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d"} err="failed to get container status \"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d\": rpc error: code = NotFound desc = could not find container \"e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d\": container with ID starting with e2e4f258ec6418552fa70691ce8ca31abb19071bc8457c1683a9feff313c387d not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.494070 4923 scope.go:117] "RemoveContainer" containerID="dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.494732 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec"} err="failed to get container status \"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec\": rpc error: code = NotFound desc = could not find container \"dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec\": container with ID starting with dff2819df0e86e3a603fa380f263b320168be8b20a28b5a27d89e1638087a0ec not found: ID does not exist" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.515786 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzchj\" (UniqueName: \"kubernetes.io/projected/dba22ce5-7738-4868-b2f2-d52f43bccbe5-kube-api-access-kzchj\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.515888 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.515938 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dba22ce5-7738-4868-b2f2-d52f43bccbe5-run-httpd\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.515974 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dba22ce5-7738-4868-b2f2-d52f43bccbe5-log-httpd\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.516013 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-config-data\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.516045 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.516089 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-scripts\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.516127 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.618389 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzchj\" (UniqueName: \"kubernetes.io/projected/dba22ce5-7738-4868-b2f2-d52f43bccbe5-kube-api-access-kzchj\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.618477 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.618504 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dba22ce5-7738-4868-b2f2-d52f43bccbe5-run-httpd\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.618539 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dba22ce5-7738-4868-b2f2-d52f43bccbe5-log-httpd\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.618561 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-config-data\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.618611 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.618645 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-scripts\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.618681 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.619663 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dba22ce5-7738-4868-b2f2-d52f43bccbe5-log-httpd\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.619961 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dba22ce5-7738-4868-b2f2-d52f43bccbe5-run-httpd\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.622317 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.622971 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.623791 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-scripts\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.624385 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.625267 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dba22ce5-7738-4868-b2f2-d52f43bccbe5-config-data\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.638168 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzchj\" (UniqueName: \"kubernetes.io/projected/dba22ce5-7738-4868-b2f2-d52f43bccbe5-kube-api-access-kzchj\") pod \"ceilometer-0\" (UID: \"dba22ce5-7738-4868-b2f2-d52f43bccbe5\") " pod="openstack/ceilometer-0" Oct 09 11:01:11 crc kubenswrapper[4923]: I1009 11:01:11.776111 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 11:01:12 crc kubenswrapper[4923]: I1009 11:01:12.258272 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 11:01:12 crc kubenswrapper[4923]: W1009 11:01:12.259082 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba22ce5_7738_4868_b2f2_d52f43bccbe5.slice/crio-8cf835efe6590a3dbf72edcd6246a40c254d5e1291b89fdc5e9226b18ac38f1c WatchSource:0}: Error finding container 8cf835efe6590a3dbf72edcd6246a40c254d5e1291b89fdc5e9226b18ac38f1c: Status 404 returned error can't find the container with id 8cf835efe6590a3dbf72edcd6246a40c254d5e1291b89fdc5e9226b18ac38f1c Oct 09 11:01:12 crc kubenswrapper[4923]: I1009 11:01:12.365271 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dba22ce5-7738-4868-b2f2-d52f43bccbe5","Type":"ContainerStarted","Data":"8cf835efe6590a3dbf72edcd6246a40c254d5e1291b89fdc5e9226b18ac38f1c"} Oct 09 11:01:12 crc kubenswrapper[4923]: I1009 11:01:12.619083 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="277b1f0f-f601-4de4-92c8-8b44b9e1da33" path="/var/lib/kubelet/pods/277b1f0f-f601-4de4-92c8-8b44b9e1da33/volumes" Oct 09 11:01:13 crc kubenswrapper[4923]: I1009 11:01:13.377986 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dba22ce5-7738-4868-b2f2-d52f43bccbe5","Type":"ContainerStarted","Data":"c63b4d8c83bc24e0992d460edf42c694d517986be3fb3eeb1275f0977ea89066"} Oct 09 11:01:13 crc kubenswrapper[4923]: I1009 11:01:13.557844 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 09 11:01:14 crc kubenswrapper[4923]: I1009 11:01:14.420878 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dba22ce5-7738-4868-b2f2-d52f43bccbe5","Type":"ContainerStarted","Data":"8812854250e3753f667449cffe38a7e64000a2c48d8fc2b30ceb7f2e9e99fa43"} Oct 09 11:01:15 crc kubenswrapper[4923]: I1009 11:01:15.432889 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 09 11:01:15 crc kubenswrapper[4923]: I1009 11:01:15.435504 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dba22ce5-7738-4868-b2f2-d52f43bccbe5","Type":"ContainerStarted","Data":"16635f852edecf030618403420b8ab1f100a9a619d8e2513b523bf069931d42e"} Oct 09 11:01:15 crc kubenswrapper[4923]: I1009 11:01:15.448314 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 09 11:01:15 crc kubenswrapper[4923]: I1009 11:01:15.507536 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 09 11:01:15 crc kubenswrapper[4923]: I1009 11:01:15.537739 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 09 11:01:16 crc kubenswrapper[4923]: I1009 11:01:16.444496 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerName="manila-scheduler" containerID="cri-o://257301b6fe39e5e129dd16f14301dc272d7380c95595a8a7e650a906874007d2" gracePeriod=30 Oct 09 11:01:16 crc kubenswrapper[4923]: I1009 11:01:16.444718 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerName="probe" containerID="cri-o://66bbb4399819f39c0d85ce3ad2fb3ae438d662ea1465ddfac6dbeacbdd7dbcdd" gracePeriod=30 Oct 09 11:01:16 crc kubenswrapper[4923]: I1009 11:01:16.445091 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerName="manila-share" containerID="cri-o://642eca6c1f47e753b16976e49881bcb864786e172f40c4a0280d0e7464ddc1dc" gracePeriod=30 Oct 09 11:01:16 crc kubenswrapper[4923]: I1009 11:01:16.445451 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerName="probe" containerID="cri-o://87189fc15435f69db9f038a33a12f39020ef7741907395ab9bacdec10d435786" gracePeriod=30 Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.458038 4923 generic.go:334] "Generic (PLEG): container finished" podID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerID="87189fc15435f69db9f038a33a12f39020ef7741907395ab9bacdec10d435786" exitCode=0 Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.458731 4923 generic.go:334] "Generic (PLEG): container finished" podID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerID="642eca6c1f47e753b16976e49881bcb864786e172f40c4a0280d0e7464ddc1dc" exitCode=1 Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.458136 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b8a91455-8e4d-45c3-8796-c4c4cd478f15","Type":"ContainerDied","Data":"87189fc15435f69db9f038a33a12f39020ef7741907395ab9bacdec10d435786"} Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.458888 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b8a91455-8e4d-45c3-8796-c4c4cd478f15","Type":"ContainerDied","Data":"642eca6c1f47e753b16976e49881bcb864786e172f40c4a0280d0e7464ddc1dc"} Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.463285 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dba22ce5-7738-4868-b2f2-d52f43bccbe5","Type":"ContainerStarted","Data":"8fc0e0c3bd5e8dda42ef1f925f660d6f357802299b2fb531b2a84bca38b7dd42"} Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.463473 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.480865 4923 generic.go:334] "Generic (PLEG): container finished" podID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerID="66bbb4399819f39c0d85ce3ad2fb3ae438d662ea1465ddfac6dbeacbdd7dbcdd" exitCode=0 Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.480947 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a8c12316-951b-48ab-a614-7a6ba90b7be8","Type":"ContainerDied","Data":"66bbb4399819f39c0d85ce3ad2fb3ae438d662ea1465ddfac6dbeacbdd7dbcdd"} Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.530462 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.567680 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.109146043 podStartE2EDuration="6.567652791s" podCreationTimestamp="2025-10-09 11:01:11 +0000 UTC" firstStartedPulling="2025-10-09 11:01:12.262171349 +0000 UTC m=+3358.330353105" lastFinishedPulling="2025-10-09 11:01:16.720678097 +0000 UTC m=+3362.788859853" observedRunningTime="2025-10-09 11:01:17.489039251 +0000 UTC m=+3363.557221327" watchObservedRunningTime="2025-10-09 11:01:17.567652791 +0000 UTC m=+3363.635834547" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.585409 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data\") pod \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.585490 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-scripts\") pod \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.585564 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-ceph\") pod \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.585715 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-combined-ca-bundle\") pod \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.585746 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-var-lib-manila\") pod \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.585831 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data-custom\") pod \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.585870 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-etc-machine-id\") pod \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.586116 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdqc6\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-kube-api-access-wdqc6\") pod \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\" (UID: \"b8a91455-8e4d-45c3-8796-c4c4cd478f15\") " Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.588101 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b8a91455-8e4d-45c3-8796-c4c4cd478f15" (UID: "b8a91455-8e4d-45c3-8796-c4c4cd478f15"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.594827 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "b8a91455-8e4d-45c3-8796-c4c4cd478f15" (UID: "b8a91455-8e4d-45c3-8796-c4c4cd478f15"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.595440 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-kube-api-access-wdqc6" (OuterVolumeSpecName: "kube-api-access-wdqc6") pod "b8a91455-8e4d-45c3-8796-c4c4cd478f15" (UID: "b8a91455-8e4d-45c3-8796-c4c4cd478f15"). InnerVolumeSpecName "kube-api-access-wdqc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.599587 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-ceph" (OuterVolumeSpecName: "ceph") pod "b8a91455-8e4d-45c3-8796-c4c4cd478f15" (UID: "b8a91455-8e4d-45c3-8796-c4c4cd478f15"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.599637 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b8a91455-8e4d-45c3-8796-c4c4cd478f15" (UID: "b8a91455-8e4d-45c3-8796-c4c4cd478f15"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.609553 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-scripts" (OuterVolumeSpecName: "scripts") pod "b8a91455-8e4d-45c3-8796-c4c4cd478f15" (UID: "b8a91455-8e4d-45c3-8796-c4c4cd478f15"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.665322 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8a91455-8e4d-45c3-8796-c4c4cd478f15" (UID: "b8a91455-8e4d-45c3-8796-c4c4cd478f15"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.689532 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.689583 4923 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-ceph\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.689595 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.689608 4923 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-var-lib-manila\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.689617 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.689642 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b8a91455-8e4d-45c3-8796-c4c4cd478f15-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.689651 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdqc6\" (UniqueName: \"kubernetes.io/projected/b8a91455-8e4d-45c3-8796-c4c4cd478f15-kube-api-access-wdqc6\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.735038 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data" (OuterVolumeSpecName: "config-data") pod "b8a91455-8e4d-45c3-8796-c4c4cd478f15" (UID: "b8a91455-8e4d-45c3-8796-c4c4cd478f15"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:17 crc kubenswrapper[4923]: I1009 11:01:17.793294 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8a91455-8e4d-45c3-8796-c4c4cd478f15-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.493801 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.508684 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b8a91455-8e4d-45c3-8796-c4c4cd478f15","Type":"ContainerDied","Data":"6516d288117a1f53ca55e64852800703c89803a252021b245ffc6229da1396c1"} Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.508775 4923 scope.go:117] "RemoveContainer" containerID="87189fc15435f69db9f038a33a12f39020ef7741907395ab9bacdec10d435786" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.551251 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.559702 4923 scope.go:117] "RemoveContainer" containerID="642eca6c1f47e753b16976e49881bcb864786e172f40c4a0280d0e7464ddc1dc" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.566875 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.582801 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Oct 09 11:01:18 crc kubenswrapper[4923]: E1009 11:01:18.583379 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerName="manila-share" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.583414 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerName="manila-share" Oct 09 11:01:18 crc kubenswrapper[4923]: E1009 11:01:18.583446 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerName="probe" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.583455 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerName="probe" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.585841 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerName="manila-share" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.585878 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" containerName="probe" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.587149 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.589703 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.620636 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8a91455-8e4d-45c3-8796-c4c4cd478f15" path="/var/lib/kubelet/pods/b8a91455-8e4d-45c3-8796-c4c4cd478f15/volumes" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.621471 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.717890 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.718061 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7fb17c25-9276-4577-b9c1-56ca34a4dece-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.718101 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-scripts\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.718130 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.718165 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bh6vt\" (UniqueName: \"kubernetes.io/projected/7fb17c25-9276-4577-b9c1-56ca34a4dece-kube-api-access-bh6vt\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.718208 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-config-data\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.718972 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7fb17c25-9276-4577-b9c1-56ca34a4dece-ceph\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.719220 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7fb17c25-9276-4577-b9c1-56ca34a4dece-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.821454 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7fb17c25-9276-4577-b9c1-56ca34a4dece-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.821585 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.821682 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7fb17c25-9276-4577-b9c1-56ca34a4dece-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.821686 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7fb17c25-9276-4577-b9c1-56ca34a4dece-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.821714 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-scripts\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.821765 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7fb17c25-9276-4577-b9c1-56ca34a4dece-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.821883 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.821919 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bh6vt\" (UniqueName: \"kubernetes.io/projected/7fb17c25-9276-4577-b9c1-56ca34a4dece-kube-api-access-bh6vt\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.821965 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-config-data\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.822051 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7fb17c25-9276-4577-b9c1-56ca34a4dece-ceph\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.826905 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-scripts\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.827099 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.827150 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-config-data\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.836272 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7fb17c25-9276-4577-b9c1-56ca34a4dece-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.837150 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7fb17c25-9276-4577-b9c1-56ca34a4dece-ceph\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.843954 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bh6vt\" (UniqueName: \"kubernetes.io/projected/7fb17c25-9276-4577-b9c1-56ca34a4dece-kube-api-access-bh6vt\") pod \"manila-share-share1-0\" (UID: \"7fb17c25-9276-4577-b9c1-56ca34a4dece\") " pod="openstack/manila-share-share1-0" Oct 09 11:01:18 crc kubenswrapper[4923]: I1009 11:01:18.953881 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Oct 09 11:01:19 crc kubenswrapper[4923]: I1009 11:01:19.568999 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.563369 4923 generic.go:334] "Generic (PLEG): container finished" podID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerID="257301b6fe39e5e129dd16f14301dc272d7380c95595a8a7e650a906874007d2" exitCode=0 Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.567824 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a8c12316-951b-48ab-a614-7a6ba90b7be8","Type":"ContainerDied","Data":"257301b6fe39e5e129dd16f14301dc272d7380c95595a8a7e650a906874007d2"} Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.572342 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7fb17c25-9276-4577-b9c1-56ca34a4dece","Type":"ContainerStarted","Data":"e94315c3006c303ef528f1fcde03979da59d6eb1c4561c569b8737a0d953c4ff"} Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.572425 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7fb17c25-9276-4577-b9c1-56ca34a4dece","Type":"ContainerStarted","Data":"6ee65281630fdc02e1253ab9640582d7a746a609d9b33721fa7dc33402225632"} Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.671517 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.777136 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-combined-ca-bundle\") pod \"a8c12316-951b-48ab-a614-7a6ba90b7be8\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.777217 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-scripts\") pod \"a8c12316-951b-48ab-a614-7a6ba90b7be8\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.777332 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55zht\" (UniqueName: \"kubernetes.io/projected/a8c12316-951b-48ab-a614-7a6ba90b7be8-kube-api-access-55zht\") pod \"a8c12316-951b-48ab-a614-7a6ba90b7be8\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.777398 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8c12316-951b-48ab-a614-7a6ba90b7be8-etc-machine-id\") pod \"a8c12316-951b-48ab-a614-7a6ba90b7be8\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.777500 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data-custom\") pod \"a8c12316-951b-48ab-a614-7a6ba90b7be8\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.777684 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data\") pod \"a8c12316-951b-48ab-a614-7a6ba90b7be8\" (UID: \"a8c12316-951b-48ab-a614-7a6ba90b7be8\") " Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.777735 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a8c12316-951b-48ab-a614-7a6ba90b7be8-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "a8c12316-951b-48ab-a614-7a6ba90b7be8" (UID: "a8c12316-951b-48ab-a614-7a6ba90b7be8"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.778306 4923 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a8c12316-951b-48ab-a614-7a6ba90b7be8-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.785988 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-scripts" (OuterVolumeSpecName: "scripts") pod "a8c12316-951b-48ab-a614-7a6ba90b7be8" (UID: "a8c12316-951b-48ab-a614-7a6ba90b7be8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.786027 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8c12316-951b-48ab-a614-7a6ba90b7be8-kube-api-access-55zht" (OuterVolumeSpecName: "kube-api-access-55zht") pod "a8c12316-951b-48ab-a614-7a6ba90b7be8" (UID: "a8c12316-951b-48ab-a614-7a6ba90b7be8"). InnerVolumeSpecName "kube-api-access-55zht". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.786967 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a8c12316-951b-48ab-a614-7a6ba90b7be8" (UID: "a8c12316-951b-48ab-a614-7a6ba90b7be8"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.851218 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a8c12316-951b-48ab-a614-7a6ba90b7be8" (UID: "a8c12316-951b-48ab-a614-7a6ba90b7be8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.882467 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55zht\" (UniqueName: \"kubernetes.io/projected/a8c12316-951b-48ab-a614-7a6ba90b7be8-kube-api-access-55zht\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.882505 4923 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.882514 4923 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.882525 4923 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.893873 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data" (OuterVolumeSpecName: "config-data") pod "a8c12316-951b-48ab-a614-7a6ba90b7be8" (UID: "a8c12316-951b-48ab-a614-7a6ba90b7be8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 11:01:20 crc kubenswrapper[4923]: I1009 11:01:20.985184 4923 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a8c12316-951b-48ab-a614-7a6ba90b7be8-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.087722 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.587387 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"a8c12316-951b-48ab-a614-7a6ba90b7be8","Type":"ContainerDied","Data":"64d2e18c94afc2ef0dc80dbb0f4bc834dcf39d11f46c9a1397f7468d108c41f4"} Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.588315 4923 scope.go:117] "RemoveContainer" containerID="66bbb4399819f39c0d85ce3ad2fb3ae438d662ea1465ddfac6dbeacbdd7dbcdd" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.587728 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.590456 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7fb17c25-9276-4577-b9c1-56ca34a4dece","Type":"ContainerStarted","Data":"1c464e2490e2c7c5b57e0ab54b5ca033387148d2ccb51efef5cc231c1b6f01b2"} Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.623811 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.623787663 podStartE2EDuration="3.623787663s" podCreationTimestamp="2025-10-09 11:01:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 11:01:21.621600482 +0000 UTC m=+3367.689782248" watchObservedRunningTime="2025-10-09 11:01:21.623787663 +0000 UTC m=+3367.691969419" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.624360 4923 scope.go:117] "RemoveContainer" containerID="257301b6fe39e5e129dd16f14301dc272d7380c95595a8a7e650a906874007d2" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.654347 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.678108 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.687761 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Oct 09 11:01:21 crc kubenswrapper[4923]: E1009 11:01:21.688382 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerName="probe" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.688401 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerName="probe" Oct 09 11:01:21 crc kubenswrapper[4923]: E1009 11:01:21.688418 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerName="manila-scheduler" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.688427 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerName="manila-scheduler" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.688605 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerName="probe" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.688634 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8c12316-951b-48ab-a614-7a6ba90b7be8" containerName="manila-scheduler" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.689837 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.709569 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.712877 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.802697 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38e3cad2-d7c0-463e-a63c-a197183d870b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.803207 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-config-data\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.803504 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs6r4\" (UniqueName: \"kubernetes.io/projected/38e3cad2-d7c0-463e-a63c-a197183d870b-kube-api-access-qs6r4\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.803688 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.803813 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.803939 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-scripts\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.906390 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-config-data\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.906467 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs6r4\" (UniqueName: \"kubernetes.io/projected/38e3cad2-d7c0-463e-a63c-a197183d870b-kube-api-access-qs6r4\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.906503 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.906535 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.906566 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-scripts\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.906618 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38e3cad2-d7c0-463e-a63c-a197183d870b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.906743 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/38e3cad2-d7c0-463e-a63c-a197183d870b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.913334 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.913893 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-config-data\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.927497 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs6r4\" (UniqueName: \"kubernetes.io/projected/38e3cad2-d7c0-463e-a63c-a197183d870b-kube-api-access-qs6r4\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.928361 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-scripts\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:21 crc kubenswrapper[4923]: I1009 11:01:21.933325 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e3cad2-d7c0-463e-a63c-a197183d870b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"38e3cad2-d7c0-463e-a63c-a197183d870b\") " pod="openstack/manila-scheduler-0" Oct 09 11:01:22 crc kubenswrapper[4923]: I1009 11:01:22.033128 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Oct 09 11:01:22 crc kubenswrapper[4923]: I1009 11:01:22.545274 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Oct 09 11:01:22 crc kubenswrapper[4923]: I1009 11:01:22.612615 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a8c12316-951b-48ab-a614-7a6ba90b7be8" path="/var/lib/kubelet/pods/a8c12316-951b-48ab-a614-7a6ba90b7be8/volumes" Oct 09 11:01:22 crc kubenswrapper[4923]: I1009 11:01:22.613841 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"38e3cad2-d7c0-463e-a63c-a197183d870b","Type":"ContainerStarted","Data":"64459c8a0e02476d4a82ec74f46bfe8572fb2001352eb35464505555b347c477"} Oct 09 11:01:23 crc kubenswrapper[4923]: I1009 11:01:23.629254 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"38e3cad2-d7c0-463e-a63c-a197183d870b","Type":"ContainerStarted","Data":"e1c0e1c6d57bf6e615f3aa6fc7dc8d722717d183e61e9402069dd40c3c7f0261"} Oct 09 11:01:23 crc kubenswrapper[4923]: I1009 11:01:23.629610 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"38e3cad2-d7c0-463e-a63c-a197183d870b","Type":"ContainerStarted","Data":"5bde4ace0e03b4a9a7f8ffd6e88e99fd9777ea4d3595a6980d2ec20d05659057"} Oct 09 11:01:23 crc kubenswrapper[4923]: I1009 11:01:23.659428 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=2.659397922 podStartE2EDuration="2.659397922s" podCreationTimestamp="2025-10-09 11:01:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 11:01:23.653008525 +0000 UTC m=+3369.721190301" watchObservedRunningTime="2025-10-09 11:01:23.659397922 +0000 UTC m=+3369.727579678" Oct 09 11:01:24 crc kubenswrapper[4923]: I1009 11:01:24.599331 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 11:01:24 crc kubenswrapper[4923]: I1009 11:01:24.599807 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 11:01:28 crc kubenswrapper[4923]: I1009 11:01:28.954722 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Oct 09 11:01:32 crc kubenswrapper[4923]: I1009 11:01:32.034851 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Oct 09 11:01:33 crc kubenswrapper[4923]: I1009 11:01:33.760659 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Oct 09 11:01:40 crc kubenswrapper[4923]: I1009 11:01:40.564026 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Oct 09 11:01:41 crc kubenswrapper[4923]: I1009 11:01:41.785026 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 09 11:01:54 crc kubenswrapper[4923]: I1009 11:01:54.600032 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 11:01:54 crc kubenswrapper[4923]: I1009 11:01:54.600879 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 11:01:54 crc kubenswrapper[4923]: I1009 11:01:54.601046 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 11:01:54 crc kubenswrapper[4923]: I1009 11:01:54.613743 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c284c9aa97a8197efff9350f0a239489485f6b80d76c31d68671f5e5bb2b8a3"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 11:01:54 crc kubenswrapper[4923]: I1009 11:01:54.613910 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://0c284c9aa97a8197efff9350f0a239489485f6b80d76c31d68671f5e5bb2b8a3" gracePeriod=600 Oct 09 11:01:54 crc kubenswrapper[4923]: I1009 11:01:54.983068 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="0c284c9aa97a8197efff9350f0a239489485f6b80d76c31d68671f5e5bb2b8a3" exitCode=0 Oct 09 11:01:54 crc kubenswrapper[4923]: I1009 11:01:54.983150 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"0c284c9aa97a8197efff9350f0a239489485f6b80d76c31d68671f5e5bb2b8a3"} Oct 09 11:01:54 crc kubenswrapper[4923]: I1009 11:01:54.983456 4923 scope.go:117] "RemoveContainer" containerID="55c8d47c6587c38a9674374cc8aa0b2e69c0c077e703ba468ca5e466d9604735" Oct 09 11:01:55 crc kubenswrapper[4923]: I1009 11:01:55.996585 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397"} Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.042589 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzzg"] Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.045730 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.058873 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzzg"] Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.106587 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-catalog-content\") pod \"redhat-marketplace-4vzzg\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.106676 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q46ml\" (UniqueName: \"kubernetes.io/projected/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-kube-api-access-q46ml\") pod \"redhat-marketplace-4vzzg\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.106776 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-utilities\") pod \"redhat-marketplace-4vzzg\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.209970 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-catalog-content\") pod \"redhat-marketplace-4vzzg\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.210856 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q46ml\" (UniqueName: \"kubernetes.io/projected/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-kube-api-access-q46ml\") pod \"redhat-marketplace-4vzzg\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.211589 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-utilities\") pod \"redhat-marketplace-4vzzg\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.211099 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-catalog-content\") pod \"redhat-marketplace-4vzzg\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.212220 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-utilities\") pod \"redhat-marketplace-4vzzg\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.237417 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q46ml\" (UniqueName: \"kubernetes.io/projected/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-kube-api-access-q46ml\") pod \"redhat-marketplace-4vzzg\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.424296 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:06 crc kubenswrapper[4923]: I1009 11:02:06.966548 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzzg"] Oct 09 11:02:07 crc kubenswrapper[4923]: I1009 11:02:07.128126 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzzg" event={"ID":"3fda9721-295b-425b-b9cc-d5bebd9c9e9e","Type":"ContainerStarted","Data":"34028d5bf2ffa748816f89474cbd03c6b2f3fff7cd7881d082d289f429e2d5f9"} Oct 09 11:02:08 crc kubenswrapper[4923]: I1009 11:02:08.144973 4923 generic.go:334] "Generic (PLEG): container finished" podID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerID="e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7" exitCode=0 Oct 09 11:02:08 crc kubenswrapper[4923]: I1009 11:02:08.145306 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzzg" event={"ID":"3fda9721-295b-425b-b9cc-d5bebd9c9e9e","Type":"ContainerDied","Data":"e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7"} Oct 09 11:02:08 crc kubenswrapper[4923]: I1009 11:02:08.148610 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 11:02:09 crc kubenswrapper[4923]: I1009 11:02:09.158378 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzzg" event={"ID":"3fda9721-295b-425b-b9cc-d5bebd9c9e9e","Type":"ContainerStarted","Data":"8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a"} Oct 09 11:02:10 crc kubenswrapper[4923]: I1009 11:02:10.171453 4923 generic.go:334] "Generic (PLEG): container finished" podID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerID="8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a" exitCode=0 Oct 09 11:02:10 crc kubenswrapper[4923]: I1009 11:02:10.171518 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzzg" event={"ID":"3fda9721-295b-425b-b9cc-d5bebd9c9e9e","Type":"ContainerDied","Data":"8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a"} Oct 09 11:02:11 crc kubenswrapper[4923]: I1009 11:02:11.185534 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzzg" event={"ID":"3fda9721-295b-425b-b9cc-d5bebd9c9e9e","Type":"ContainerStarted","Data":"da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd"} Oct 09 11:02:11 crc kubenswrapper[4923]: I1009 11:02:11.218569 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4vzzg" podStartSLOduration=2.758984028 podStartE2EDuration="5.218543382s" podCreationTimestamp="2025-10-09 11:02:06 +0000 UTC" firstStartedPulling="2025-10-09 11:02:08.148340027 +0000 UTC m=+3414.216521783" lastFinishedPulling="2025-10-09 11:02:10.607899371 +0000 UTC m=+3416.676081137" observedRunningTime="2025-10-09 11:02:11.213534034 +0000 UTC m=+3417.281715810" watchObservedRunningTime="2025-10-09 11:02:11.218543382 +0000 UTC m=+3417.286725138" Oct 09 11:02:16 crc kubenswrapper[4923]: I1009 11:02:16.425143 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:16 crc kubenswrapper[4923]: I1009 11:02:16.425825 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:16 crc kubenswrapper[4923]: I1009 11:02:16.481427 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:17 crc kubenswrapper[4923]: I1009 11:02:17.300085 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:17 crc kubenswrapper[4923]: I1009 11:02:17.353302 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzzg"] Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.266158 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-4vzzg" podUID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerName="registry-server" containerID="cri-o://da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd" gracePeriod=2 Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.783620 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.846741 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-catalog-content\") pod \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.847317 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q46ml\" (UniqueName: \"kubernetes.io/projected/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-kube-api-access-q46ml\") pod \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.847643 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-utilities\") pod \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\" (UID: \"3fda9721-295b-425b-b9cc-d5bebd9c9e9e\") " Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.848777 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-utilities" (OuterVolumeSpecName: "utilities") pod "3fda9721-295b-425b-b9cc-d5bebd9c9e9e" (UID: "3fda9721-295b-425b-b9cc-d5bebd9c9e9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.850106 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.856085 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-kube-api-access-q46ml" (OuterVolumeSpecName: "kube-api-access-q46ml") pod "3fda9721-295b-425b-b9cc-d5bebd9c9e9e" (UID: "3fda9721-295b-425b-b9cc-d5bebd9c9e9e"). InnerVolumeSpecName "kube-api-access-q46ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.860724 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3fda9721-295b-425b-b9cc-d5bebd9c9e9e" (UID: "3fda9721-295b-425b-b9cc-d5bebd9c9e9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.952503 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 11:02:19 crc kubenswrapper[4923]: I1009 11:02:19.952541 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q46ml\" (UniqueName: \"kubernetes.io/projected/3fda9721-295b-425b-b9cc-d5bebd9c9e9e-kube-api-access-q46ml\") on node \"crc\" DevicePath \"\"" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.279065 4923 generic.go:334] "Generic (PLEG): container finished" podID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerID="da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd" exitCode=0 Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.279179 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzzg" event={"ID":"3fda9721-295b-425b-b9cc-d5bebd9c9e9e","Type":"ContainerDied","Data":"da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd"} Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.279580 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4vzzg" event={"ID":"3fda9721-295b-425b-b9cc-d5bebd9c9e9e","Type":"ContainerDied","Data":"34028d5bf2ffa748816f89474cbd03c6b2f3fff7cd7881d082d289f429e2d5f9"} Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.279611 4923 scope.go:117] "RemoveContainer" containerID="da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.279248 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4vzzg" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.323147 4923 scope.go:117] "RemoveContainer" containerID="8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.331031 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzzg"] Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.352305 4923 scope.go:117] "RemoveContainer" containerID="e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.364805 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-4vzzg"] Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.403792 4923 scope.go:117] "RemoveContainer" containerID="da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd" Oct 09 11:02:20 crc kubenswrapper[4923]: E1009 11:02:20.404449 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd\": container with ID starting with da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd not found: ID does not exist" containerID="da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.404484 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd"} err="failed to get container status \"da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd\": rpc error: code = NotFound desc = could not find container \"da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd\": container with ID starting with da7141ddf1f52b322c3f3accf50e3dcd84ed75147fa9a0e0082af20ece5600dd not found: ID does not exist" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.404521 4923 scope.go:117] "RemoveContainer" containerID="8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a" Oct 09 11:02:20 crc kubenswrapper[4923]: E1009 11:02:20.405036 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a\": container with ID starting with 8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a not found: ID does not exist" containerID="8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.405065 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a"} err="failed to get container status \"8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a\": rpc error: code = NotFound desc = could not find container \"8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a\": container with ID starting with 8edca28f7c9f4fddeb5bb09736d73e90ebfd76dd62d9d3eadb013aee71c2435a not found: ID does not exist" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.405082 4923 scope.go:117] "RemoveContainer" containerID="e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7" Oct 09 11:02:20 crc kubenswrapper[4923]: E1009 11:02:20.405389 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7\": container with ID starting with e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7 not found: ID does not exist" containerID="e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.405422 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7"} err="failed to get container status \"e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7\": rpc error: code = NotFound desc = could not find container \"e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7\": container with ID starting with e52ccc167c0dca16aa153c393b321954a8950a1199503f024845aab86c625ad7 not found: ID does not exist" Oct 09 11:02:20 crc kubenswrapper[4923]: I1009 11:02:20.615591 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" path="/var/lib/kubelet/pods/3fda9721-295b-425b-b9cc-d5bebd9c9e9e/volumes" Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.754442 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj"] Oct 09 11:02:41 crc kubenswrapper[4923]: E1009 11:02:41.755622 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerName="extract-utilities" Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.755641 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerName="extract-utilities" Oct 09 11:02:41 crc kubenswrapper[4923]: E1009 11:02:41.755672 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerName="extract-content" Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.755680 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerName="extract-content" Oct 09 11:02:41 crc kubenswrapper[4923]: E1009 11:02:41.755697 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerName="registry-server" Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.755706 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerName="registry-server" Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.755978 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="3fda9721-295b-425b-b9cc-d5bebd9c9e9e" containerName="registry-server" Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.757280 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.785717 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj"] Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.795945 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsnwv\" (UniqueName: \"kubernetes.io/projected/b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0-kube-api-access-vsnwv\") pod \"openstack-operator-controller-operator-5c7bf7948-jnnfj\" (UID: \"b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0\") " pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.898851 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsnwv\" (UniqueName: \"kubernetes.io/projected/b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0-kube-api-access-vsnwv\") pod \"openstack-operator-controller-operator-5c7bf7948-jnnfj\" (UID: \"b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0\") " pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" Oct 09 11:02:41 crc kubenswrapper[4923]: I1009 11:02:41.935006 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsnwv\" (UniqueName: \"kubernetes.io/projected/b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0-kube-api-access-vsnwv\") pod \"openstack-operator-controller-operator-5c7bf7948-jnnfj\" (UID: \"b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0\") " pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" Oct 09 11:02:42 crc kubenswrapper[4923]: I1009 11:02:42.079372 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" Oct 09 11:02:42 crc kubenswrapper[4923]: I1009 11:02:42.722511 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj"] Oct 09 11:02:43 crc kubenswrapper[4923]: I1009 11:02:43.521050 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" event={"ID":"b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0","Type":"ContainerStarted","Data":"f1d8c0507938de6f392e69a6a6d3c0f1c9ad0d6f85bc4190e2dd26cddcbe8832"} Oct 09 11:02:43 crc kubenswrapper[4923]: I1009 11:02:43.521654 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" Oct 09 11:02:43 crc kubenswrapper[4923]: I1009 11:02:43.521673 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" event={"ID":"b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0","Type":"ContainerStarted","Data":"6bea5087b88d4e841efa7f3f75b5e40f75eb86adc8dd2ec856a1bfbc5d00e4a6"} Oct 09 11:02:43 crc kubenswrapper[4923]: I1009 11:02:43.521687 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" event={"ID":"b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0","Type":"ContainerStarted","Data":"b5cfa4e8ee4b36e5a87a00268312a209a0b199c901b5934ba54ede6f8db41c99"} Oct 09 11:02:43 crc kubenswrapper[4923]: I1009 11:02:43.568566 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" podStartSLOduration=2.568542289 podStartE2EDuration="2.568542289s" podCreationTimestamp="2025-10-09 11:02:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 11:02:43.56099232 +0000 UTC m=+3449.629174086" watchObservedRunningTime="2025-10-09 11:02:43.568542289 +0000 UTC m=+3449.636724045" Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.083214 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5c7bf7948-jnnfj" Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.167079 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4"] Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.167422 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" podUID="9aeb5354-727c-44fa-85dc-4921a593090f" containerName="operator" containerID="cri-o://ae749477ba39ea025dd6162b3445b514c0bed570c20bb974f7fd8a031e5efa5f" gracePeriod=10 Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.167585 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" podUID="9aeb5354-727c-44fa-85dc-4921a593090f" containerName="kube-rbac-proxy" containerID="cri-o://09f2086bcc2d20707e039c3760222d4f5bbd8c8fc2dc6c7d63e16e401e3f86f5" gracePeriod=10 Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.606827 4923 generic.go:334] "Generic (PLEG): container finished" podID="9aeb5354-727c-44fa-85dc-4921a593090f" containerID="09f2086bcc2d20707e039c3760222d4f5bbd8c8fc2dc6c7d63e16e401e3f86f5" exitCode=0 Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.607304 4923 generic.go:334] "Generic (PLEG): container finished" podID="9aeb5354-727c-44fa-85dc-4921a593090f" containerID="ae749477ba39ea025dd6162b3445b514c0bed570c20bb974f7fd8a031e5efa5f" exitCode=0 Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.616328 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" event={"ID":"9aeb5354-727c-44fa-85dc-4921a593090f","Type":"ContainerDied","Data":"09f2086bcc2d20707e039c3760222d4f5bbd8c8fc2dc6c7d63e16e401e3f86f5"} Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.616393 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" event={"ID":"9aeb5354-727c-44fa-85dc-4921a593090f","Type":"ContainerDied","Data":"ae749477ba39ea025dd6162b3445b514c0bed570c20bb974f7fd8a031e5efa5f"} Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.616410 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" event={"ID":"9aeb5354-727c-44fa-85dc-4921a593090f","Type":"ContainerDied","Data":"085948d80608a446f03317e618ace39e81ba2c8fc0cac433c78f688a9cfa3763"} Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.616425 4923 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="085948d80608a446f03317e618ace39e81ba2c8fc0cac433c78f688a9cfa3763" Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.695507 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.778432 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7x9n\" (UniqueName: \"kubernetes.io/projected/9aeb5354-727c-44fa-85dc-4921a593090f-kube-api-access-z7x9n\") pod \"9aeb5354-727c-44fa-85dc-4921a593090f\" (UID: \"9aeb5354-727c-44fa-85dc-4921a593090f\") " Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.788756 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aeb5354-727c-44fa-85dc-4921a593090f-kube-api-access-z7x9n" (OuterVolumeSpecName: "kube-api-access-z7x9n") pod "9aeb5354-727c-44fa-85dc-4921a593090f" (UID: "9aeb5354-727c-44fa-85dc-4921a593090f"). InnerVolumeSpecName "kube-api-access-z7x9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:02:52 crc kubenswrapper[4923]: I1009 11:02:52.881945 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7x9n\" (UniqueName: \"kubernetes.io/projected/9aeb5354-727c-44fa-85dc-4921a593090f-kube-api-access-z7x9n\") on node \"crc\" DevicePath \"\"" Oct 09 11:02:53 crc kubenswrapper[4923]: I1009 11:02:53.623559 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4" Oct 09 11:02:53 crc kubenswrapper[4923]: I1009 11:02:53.673062 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4"] Oct 09 11:02:53 crc kubenswrapper[4923]: I1009 11:02:53.688105 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-848c57cb5c-f88g4"] Oct 09 11:02:54 crc kubenswrapper[4923]: I1009 11:02:54.612656 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9aeb5354-727c-44fa-85dc-4921a593090f" path="/var/lib/kubelet/pods/9aeb5354-727c-44fa-85dc-4921a593090f/volumes" Oct 09 11:03:20 crc kubenswrapper[4923]: I1009 11:03:20.645042 4923 scope.go:117] "RemoveContainer" containerID="ae749477ba39ea025dd6162b3445b514c0bed570c20bb974f7fd8a031e5efa5f" Oct 09 11:03:20 crc kubenswrapper[4923]: I1009 11:03:20.677152 4923 scope.go:117] "RemoveContainer" containerID="09f2086bcc2d20707e039c3760222d4f5bbd8c8fc2dc6c7d63e16e401e3f86f5" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.001344 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-776875b78c-5975g"] Oct 09 11:03:29 crc kubenswrapper[4923]: E1009 11:03:29.002938 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aeb5354-727c-44fa-85dc-4921a593090f" containerName="kube-rbac-proxy" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.002964 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aeb5354-727c-44fa-85dc-4921a593090f" containerName="kube-rbac-proxy" Oct 09 11:03:29 crc kubenswrapper[4923]: E1009 11:03:29.002993 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aeb5354-727c-44fa-85dc-4921a593090f" containerName="operator" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.003001 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aeb5354-727c-44fa-85dc-4921a593090f" containerName="operator" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.003303 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aeb5354-727c-44fa-85dc-4921a593090f" containerName="kube-rbac-proxy" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.003332 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aeb5354-727c-44fa-85dc-4921a593090f" containerName="operator" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.004904 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.028117 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-776875b78c-5975g"] Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.192213 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjrmd\" (UniqueName: \"kubernetes.io/projected/c703820f-a08c-4666-abc7-48954e40d927-kube-api-access-kjrmd\") pod \"test-operator-controller-manager-776875b78c-5975g\" (UID: \"c703820f-a08c-4666-abc7-48954e40d927\") " pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.295638 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjrmd\" (UniqueName: \"kubernetes.io/projected/c703820f-a08c-4666-abc7-48954e40d927-kube-api-access-kjrmd\") pod \"test-operator-controller-manager-776875b78c-5975g\" (UID: \"c703820f-a08c-4666-abc7-48954e40d927\") " pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.322096 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjrmd\" (UniqueName: \"kubernetes.io/projected/c703820f-a08c-4666-abc7-48954e40d927-kube-api-access-kjrmd\") pod \"test-operator-controller-manager-776875b78c-5975g\" (UID: \"c703820f-a08c-4666-abc7-48954e40d927\") " pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.343055 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:03:29 crc kubenswrapper[4923]: I1009 11:03:29.848204 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-776875b78c-5975g"] Oct 09 11:03:30 crc kubenswrapper[4923]: I1009 11:03:30.010013 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerStarted","Data":"ed8470463254556aabdf7e84acec2411e93284c9459ac0c721fe6804888ba526"} Oct 09 11:03:32 crc kubenswrapper[4923]: I1009 11:03:32.031811 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerStarted","Data":"f67fc372179fc317c125c19015903d84fafdeff85ac4138c2295348dab1cd413"} Oct 09 11:03:32 crc kubenswrapper[4923]: I1009 11:03:32.032657 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:03:32 crc kubenswrapper[4923]: I1009 11:03:32.032671 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerStarted","Data":"9db60a6471e24ee9345cd8dfc83b94aa0a95eff63f1d886e992291d3e879ea8a"} Oct 09 11:03:32 crc kubenswrapper[4923]: I1009 11:03:32.051225 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" podStartSLOduration=2.990091453 podStartE2EDuration="4.051200504s" podCreationTimestamp="2025-10-09 11:03:28 +0000 UTC" firstStartedPulling="2025-10-09 11:03:29.856737069 +0000 UTC m=+3495.924918825" lastFinishedPulling="2025-10-09 11:03:30.9178461 +0000 UTC m=+3496.986027876" observedRunningTime="2025-10-09 11:03:32.048864519 +0000 UTC m=+3498.117046275" watchObservedRunningTime="2025-10-09 11:03:32.051200504 +0000 UTC m=+3498.119382260" Oct 09 11:03:39 crc kubenswrapper[4923]: I1009 11:03:39.346214 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:03:39 crc kubenswrapper[4923]: I1009 11:03:39.405278 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-flh7l"] Oct 09 11:03:39 crc kubenswrapper[4923]: I1009 11:03:39.405584 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" podUID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerName="manager" containerID="cri-o://32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60" gracePeriod=10 Oct 09 11:03:39 crc kubenswrapper[4923]: I1009 11:03:39.405922 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" podUID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerName="kube-rbac-proxy" containerID="cri-o://e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd" gracePeriod=10 Oct 09 11:03:39 crc kubenswrapper[4923]: I1009 11:03:39.902270 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" Oct 09 11:03:39 crc kubenswrapper[4923]: I1009 11:03:39.997128 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrwvc\" (UniqueName: \"kubernetes.io/projected/f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8-kube-api-access-lrwvc\") pod \"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8\" (UID: \"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8\") " Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.026431 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8-kube-api-access-lrwvc" (OuterVolumeSpecName: "kube-api-access-lrwvc") pod "f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" (UID: "f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8"). InnerVolumeSpecName "kube-api-access-lrwvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.100488 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrwvc\" (UniqueName: \"kubernetes.io/projected/f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8-kube-api-access-lrwvc\") on node \"crc\" DevicePath \"\"" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.130856 4923 generic.go:334] "Generic (PLEG): container finished" podID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerID="e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd" exitCode=0 Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.130901 4923 generic.go:334] "Generic (PLEG): container finished" podID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerID="32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60" exitCode=0 Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.130931 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" event={"ID":"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8","Type":"ContainerDied","Data":"e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd"} Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.130966 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.130972 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" event={"ID":"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8","Type":"ContainerDied","Data":"32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60"} Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.131316 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5458f77c4-flh7l" event={"ID":"f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8","Type":"ContainerDied","Data":"d77859ab34b9a08bbc22b0e8b5719ecbd7bca4fe1b987fe715a3969059a3e909"} Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.130990 4923 scope.go:117] "RemoveContainer" containerID="e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.184663 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-flh7l"] Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.195720 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/test-operator-controller-manager-5458f77c4-flh7l"] Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.370409 4923 scope.go:117] "RemoveContainer" containerID="32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.568521 4923 scope.go:117] "RemoveContainer" containerID="e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd" Oct 09 11:03:40 crc kubenswrapper[4923]: E1009 11:03:40.568988 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd\": container with ID starting with e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd not found: ID does not exist" containerID="e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.569023 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd"} err="failed to get container status \"e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd\": rpc error: code = NotFound desc = could not find container \"e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd\": container with ID starting with e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd not found: ID does not exist" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.569059 4923 scope.go:117] "RemoveContainer" containerID="32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60" Oct 09 11:03:40 crc kubenswrapper[4923]: E1009 11:03:40.569738 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60\": container with ID starting with 32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60 not found: ID does not exist" containerID="32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.569829 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60"} err="failed to get container status \"32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60\": rpc error: code = NotFound desc = could not find container \"32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60\": container with ID starting with 32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60 not found: ID does not exist" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.569866 4923 scope.go:117] "RemoveContainer" containerID="e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.570292 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd"} err="failed to get container status \"e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd\": rpc error: code = NotFound desc = could not find container \"e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd\": container with ID starting with e8de43030d95b5e1eb9f2b9fef81f573d27e2958b24802e7b453d952fcf3e6bd not found: ID does not exist" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.570359 4923 scope.go:117] "RemoveContainer" containerID="32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.570734 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60"} err="failed to get container status \"32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60\": rpc error: code = NotFound desc = could not find container \"32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60\": container with ID starting with 32aa79d952bc203e49c89ae2b0f3824a1beee5419c7618aa52941faf6bd39b60 not found: ID does not exist" Oct 09 11:03:40 crc kubenswrapper[4923]: I1009 11:03:40.615545 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" path="/var/lib/kubelet/pods/f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8/volumes" Oct 09 11:03:54 crc kubenswrapper[4923]: I1009 11:03:54.600172 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 11:03:54 crc kubenswrapper[4923]: I1009 11:03:54.600915 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 11:04:24 crc kubenswrapper[4923]: I1009 11:04:24.599665 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 11:04:24 crc kubenswrapper[4923]: I1009 11:04:24.600560 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.265586 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x42hs"] Oct 09 11:04:44 crc kubenswrapper[4923]: E1009 11:04:44.266812 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerName="kube-rbac-proxy" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.266829 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerName="kube-rbac-proxy" Oct 09 11:04:44 crc kubenswrapper[4923]: E1009 11:04:44.266875 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerName="manager" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.266882 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerName="manager" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.267097 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerName="kube-rbac-proxy" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.267118 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f81f5f01-4b40-43ff-b3dd-fb1e70ba7fb8" containerName="manager" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.268773 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.284264 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x42hs"] Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.373370 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-utilities\") pod \"redhat-operators-x42hs\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.374019 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-catalog-content\") pod \"redhat-operators-x42hs\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.374216 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnwss\" (UniqueName: \"kubernetes.io/projected/374a3519-84ad-435a-b507-ea75996a60f0-kube-api-access-gnwss\") pod \"redhat-operators-x42hs\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.477216 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-utilities\") pod \"redhat-operators-x42hs\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.478011 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-catalog-content\") pod \"redhat-operators-x42hs\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.478118 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnwss\" (UniqueName: \"kubernetes.io/projected/374a3519-84ad-435a-b507-ea75996a60f0-kube-api-access-gnwss\") pod \"redhat-operators-x42hs\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.477888 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-utilities\") pod \"redhat-operators-x42hs\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.478972 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-catalog-content\") pod \"redhat-operators-x42hs\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.503699 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnwss\" (UniqueName: \"kubernetes.io/projected/374a3519-84ad-435a-b507-ea75996a60f0-kube-api-access-gnwss\") pod \"redhat-operators-x42hs\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:44 crc kubenswrapper[4923]: I1009 11:04:44.597707 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:45 crc kubenswrapper[4923]: I1009 11:04:45.118511 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x42hs"] Oct 09 11:04:45 crc kubenswrapper[4923]: I1009 11:04:45.818018 4923 generic.go:334] "Generic (PLEG): container finished" podID="374a3519-84ad-435a-b507-ea75996a60f0" containerID="e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd" exitCode=0 Oct 09 11:04:45 crc kubenswrapper[4923]: I1009 11:04:45.818113 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x42hs" event={"ID":"374a3519-84ad-435a-b507-ea75996a60f0","Type":"ContainerDied","Data":"e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd"} Oct 09 11:04:45 crc kubenswrapper[4923]: I1009 11:04:45.818490 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x42hs" event={"ID":"374a3519-84ad-435a-b507-ea75996a60f0","Type":"ContainerStarted","Data":"9fc5daa2bed858d11641f349bc1eb85d8ac807e67a001f674be552c4a354194e"} Oct 09 11:04:47 crc kubenswrapper[4923]: I1009 11:04:47.842552 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x42hs" event={"ID":"374a3519-84ad-435a-b507-ea75996a60f0","Type":"ContainerStarted","Data":"7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a"} Oct 09 11:04:48 crc kubenswrapper[4923]: I1009 11:04:48.855066 4923 generic.go:334] "Generic (PLEG): container finished" podID="374a3519-84ad-435a-b507-ea75996a60f0" containerID="7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a" exitCode=0 Oct 09 11:04:48 crc kubenswrapper[4923]: I1009 11:04:48.855141 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x42hs" event={"ID":"374a3519-84ad-435a-b507-ea75996a60f0","Type":"ContainerDied","Data":"7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a"} Oct 09 11:04:49 crc kubenswrapper[4923]: I1009 11:04:49.869355 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x42hs" event={"ID":"374a3519-84ad-435a-b507-ea75996a60f0","Type":"ContainerStarted","Data":"e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce"} Oct 09 11:04:49 crc kubenswrapper[4923]: I1009 11:04:49.890165 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x42hs" podStartSLOduration=2.119809042 podStartE2EDuration="5.890120938s" podCreationTimestamp="2025-10-09 11:04:44 +0000 UTC" firstStartedPulling="2025-10-09 11:04:45.821289044 +0000 UTC m=+3571.889470820" lastFinishedPulling="2025-10-09 11:04:49.59160096 +0000 UTC m=+3575.659782716" observedRunningTime="2025-10-09 11:04:49.886997071 +0000 UTC m=+3575.955178837" watchObservedRunningTime="2025-10-09 11:04:49.890120938 +0000 UTC m=+3575.958302694" Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.598693 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.599621 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.600341 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.600621 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.613893 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.614494 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.614566 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" gracePeriod=600 Oct 09 11:04:54 crc kubenswrapper[4923]: E1009 11:04:54.755298 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.955166 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" exitCode=0 Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.955244 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397"} Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.955361 4923 scope.go:117] "RemoveContainer" containerID="0c284c9aa97a8197efff9350f0a239489485f6b80d76c31d68671f5e5bb2b8a3" Oct 09 11:04:54 crc kubenswrapper[4923]: I1009 11:04:54.957837 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:04:54 crc kubenswrapper[4923]: E1009 11:04:54.958373 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:04:55 crc kubenswrapper[4923]: I1009 11:04:55.654956 4923 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x42hs" podUID="374a3519-84ad-435a-b507-ea75996a60f0" containerName="registry-server" probeResult="failure" output=< Oct 09 11:04:55 crc kubenswrapper[4923]: timeout: failed to connect service ":50051" within 1s Oct 09 11:04:55 crc kubenswrapper[4923]: > Oct 09 11:05:04 crc kubenswrapper[4923]: I1009 11:05:04.658902 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:05:04 crc kubenswrapper[4923]: I1009 11:05:04.714632 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:05:04 crc kubenswrapper[4923]: I1009 11:05:04.909884 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x42hs"] Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.064228 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x42hs" podUID="374a3519-84ad-435a-b507-ea75996a60f0" containerName="registry-server" containerID="cri-o://e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce" gracePeriod=2 Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.538317 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.602297 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:05:06 crc kubenswrapper[4923]: E1009 11:05:06.602710 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.649731 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gnwss\" (UniqueName: \"kubernetes.io/projected/374a3519-84ad-435a-b507-ea75996a60f0-kube-api-access-gnwss\") pod \"374a3519-84ad-435a-b507-ea75996a60f0\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.649824 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-utilities\") pod \"374a3519-84ad-435a-b507-ea75996a60f0\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.650083 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-catalog-content\") pod \"374a3519-84ad-435a-b507-ea75996a60f0\" (UID: \"374a3519-84ad-435a-b507-ea75996a60f0\") " Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.651134 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-utilities" (OuterVolumeSpecName: "utilities") pod "374a3519-84ad-435a-b507-ea75996a60f0" (UID: "374a3519-84ad-435a-b507-ea75996a60f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.667027 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/374a3519-84ad-435a-b507-ea75996a60f0-kube-api-access-gnwss" (OuterVolumeSpecName: "kube-api-access-gnwss") pod "374a3519-84ad-435a-b507-ea75996a60f0" (UID: "374a3519-84ad-435a-b507-ea75996a60f0"). InnerVolumeSpecName "kube-api-access-gnwss". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.748400 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "374a3519-84ad-435a-b507-ea75996a60f0" (UID: "374a3519-84ad-435a-b507-ea75996a60f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.753030 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.753073 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gnwss\" (UniqueName: \"kubernetes.io/projected/374a3519-84ad-435a-b507-ea75996a60f0-kube-api-access-gnwss\") on node \"crc\" DevicePath \"\"" Oct 09 11:05:06 crc kubenswrapper[4923]: I1009 11:05:06.753086 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/374a3519-84ad-435a-b507-ea75996a60f0-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.077072 4923 generic.go:334] "Generic (PLEG): container finished" podID="374a3519-84ad-435a-b507-ea75996a60f0" containerID="e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce" exitCode=0 Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.077212 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x42hs" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.077197 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x42hs" event={"ID":"374a3519-84ad-435a-b507-ea75996a60f0","Type":"ContainerDied","Data":"e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce"} Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.078320 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x42hs" event={"ID":"374a3519-84ad-435a-b507-ea75996a60f0","Type":"ContainerDied","Data":"9fc5daa2bed858d11641f349bc1eb85d8ac807e67a001f674be552c4a354194e"} Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.078350 4923 scope.go:117] "RemoveContainer" containerID="e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.109150 4923 scope.go:117] "RemoveContainer" containerID="7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.132201 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x42hs"] Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.144058 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x42hs"] Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.152038 4923 scope.go:117] "RemoveContainer" containerID="e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.214602 4923 scope.go:117] "RemoveContainer" containerID="e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce" Oct 09 11:05:07 crc kubenswrapper[4923]: E1009 11:05:07.215124 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce\": container with ID starting with e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce not found: ID does not exist" containerID="e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.215168 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce"} err="failed to get container status \"e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce\": rpc error: code = NotFound desc = could not find container \"e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce\": container with ID starting with e62e7a3c08fa12638ef0bf9b6db30fb5915a2ffd2a3410f98551aab11a8195ce not found: ID does not exist" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.215224 4923 scope.go:117] "RemoveContainer" containerID="7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a" Oct 09 11:05:07 crc kubenswrapper[4923]: E1009 11:05:07.215526 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a\": container with ID starting with 7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a not found: ID does not exist" containerID="7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.215556 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a"} err="failed to get container status \"7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a\": rpc error: code = NotFound desc = could not find container \"7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a\": container with ID starting with 7ae85fcc992670e7ee0b6c676684e2dbea2ca172b391e176de89224ca56cbc2a not found: ID does not exist" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.215571 4923 scope.go:117] "RemoveContainer" containerID="e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd" Oct 09 11:05:07 crc kubenswrapper[4923]: E1009 11:05:07.215884 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd\": container with ID starting with e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd not found: ID does not exist" containerID="e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd" Oct 09 11:05:07 crc kubenswrapper[4923]: I1009 11:05:07.215969 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd"} err="failed to get container status \"e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd\": rpc error: code = NotFound desc = could not find container \"e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd\": container with ID starting with e7366ade49357bb9b3964fc1bb17835c5a6a58539194e60f4b13ff4d5c1039cd not found: ID does not exist" Oct 09 11:05:08 crc kubenswrapper[4923]: I1009 11:05:08.613229 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="374a3519-84ad-435a-b507-ea75996a60f0" path="/var/lib/kubelet/pods/374a3519-84ad-435a-b507-ea75996a60f0/volumes" Oct 09 11:05:18 crc kubenswrapper[4923]: I1009 11:05:18.602383 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:05:18 crc kubenswrapper[4923]: E1009 11:05:18.604251 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:05:33 crc kubenswrapper[4923]: I1009 11:05:33.601357 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:05:33 crc kubenswrapper[4923]: E1009 11:05:33.602409 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:05:47 crc kubenswrapper[4923]: I1009 11:05:47.602613 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:05:47 crc kubenswrapper[4923]: E1009 11:05:47.604340 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:06:01 crc kubenswrapper[4923]: I1009 11:06:01.601640 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:06:01 crc kubenswrapper[4923]: E1009 11:06:01.602551 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.214293 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-p2kmf/must-gather-x5cfb"] Oct 09 11:06:10 crc kubenswrapper[4923]: E1009 11:06:10.216667 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="374a3519-84ad-435a-b507-ea75996a60f0" containerName="extract-utilities" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.216744 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="374a3519-84ad-435a-b507-ea75996a60f0" containerName="extract-utilities" Oct 09 11:06:10 crc kubenswrapper[4923]: E1009 11:06:10.216885 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="374a3519-84ad-435a-b507-ea75996a60f0" containerName="registry-server" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.216951 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="374a3519-84ad-435a-b507-ea75996a60f0" containerName="registry-server" Oct 09 11:06:10 crc kubenswrapper[4923]: E1009 11:06:10.217037 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="374a3519-84ad-435a-b507-ea75996a60f0" containerName="extract-content" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.217100 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="374a3519-84ad-435a-b507-ea75996a60f0" containerName="extract-content" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.217361 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="374a3519-84ad-435a-b507-ea75996a60f0" containerName="registry-server" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.218545 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.227475 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-p2kmf"/"openshift-service-ca.crt" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.227725 4923 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-p2kmf"/"kube-root-ca.crt" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.228130 4923 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-p2kmf"/"default-dockercfg-sxc9f" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.231702 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-p2kmf/must-gather-x5cfb"] Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.295190 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shfgd\" (UniqueName: \"kubernetes.io/projected/2db7ad3b-f634-4a87-a5af-f4d00b913e85-kube-api-access-shfgd\") pod \"must-gather-x5cfb\" (UID: \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\") " pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.295285 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2db7ad3b-f634-4a87-a5af-f4d00b913e85-must-gather-output\") pod \"must-gather-x5cfb\" (UID: \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\") " pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.398289 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2db7ad3b-f634-4a87-a5af-f4d00b913e85-must-gather-output\") pod \"must-gather-x5cfb\" (UID: \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\") " pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.398502 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shfgd\" (UniqueName: \"kubernetes.io/projected/2db7ad3b-f634-4a87-a5af-f4d00b913e85-kube-api-access-shfgd\") pod \"must-gather-x5cfb\" (UID: \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\") " pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.399006 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2db7ad3b-f634-4a87-a5af-f4d00b913e85-must-gather-output\") pod \"must-gather-x5cfb\" (UID: \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\") " pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.423157 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shfgd\" (UniqueName: \"kubernetes.io/projected/2db7ad3b-f634-4a87-a5af-f4d00b913e85-kube-api-access-shfgd\") pod \"must-gather-x5cfb\" (UID: \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\") " pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:06:10 crc kubenswrapper[4923]: I1009 11:06:10.544377 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:06:11 crc kubenswrapper[4923]: I1009 11:06:11.059084 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-p2kmf/must-gather-x5cfb"] Oct 09 11:06:11 crc kubenswrapper[4923]: I1009 11:06:11.773981 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" event={"ID":"2db7ad3b-f634-4a87-a5af-f4d00b913e85","Type":"ContainerStarted","Data":"02aa6a8cff5da64e08b03fe0b80d75adf5c85b3b9b68cf5df8d7508a599444e1"} Oct 09 11:06:12 crc kubenswrapper[4923]: I1009 11:06:12.790623 4923 generic.go:334] "Generic (PLEG): container finished" podID="c703820f-a08c-4666-abc7-48954e40d927" containerID="9db60a6471e24ee9345cd8dfc83b94aa0a95eff63f1d886e992291d3e879ea8a" exitCode=1 Oct 09 11:06:12 crc kubenswrapper[4923]: I1009 11:06:12.790710 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerDied","Data":"9db60a6471e24ee9345cd8dfc83b94aa0a95eff63f1d886e992291d3e879ea8a"} Oct 09 11:06:12 crc kubenswrapper[4923]: I1009 11:06:12.792156 4923 scope.go:117] "RemoveContainer" containerID="9db60a6471e24ee9345cd8dfc83b94aa0a95eff63f1d886e992291d3e879ea8a" Oct 09 11:06:13 crc kubenswrapper[4923]: I1009 11:06:13.812306 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerStarted","Data":"670db4a03e2a459f6e766d46d957174f0b06617bc7aeac6d8ff72e27c423a1f6"} Oct 09 11:06:13 crc kubenswrapper[4923]: I1009 11:06:13.814098 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:06:15 crc kubenswrapper[4923]: I1009 11:06:15.601962 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:06:15 crc kubenswrapper[4923]: E1009 11:06:15.602942 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:06:16 crc kubenswrapper[4923]: I1009 11:06:16.848888 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" event={"ID":"2db7ad3b-f634-4a87-a5af-f4d00b913e85","Type":"ContainerStarted","Data":"365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e"} Oct 09 11:06:17 crc kubenswrapper[4923]: I1009 11:06:17.864401 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" event={"ID":"2db7ad3b-f634-4a87-a5af-f4d00b913e85","Type":"ContainerStarted","Data":"a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319"} Oct 09 11:06:17 crc kubenswrapper[4923]: I1009 11:06:17.894003 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" podStartSLOduration=2.601390874 podStartE2EDuration="7.893977568s" podCreationTimestamp="2025-10-09 11:06:10 +0000 UTC" firstStartedPulling="2025-10-09 11:06:11.074875979 +0000 UTC m=+3657.143057735" lastFinishedPulling="2025-10-09 11:06:16.367462673 +0000 UTC m=+3662.435644429" observedRunningTime="2025-10-09 11:06:17.890300136 +0000 UTC m=+3663.958481892" watchObservedRunningTime="2025-10-09 11:06:17.893977568 +0000 UTC m=+3663.962159344" Oct 09 11:06:19 crc kubenswrapper[4923]: I1009 11:06:19.345062 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.402241 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-p2kmf/crc-debug-6sqtv"] Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.404802 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.588359 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8q6p\" (UniqueName: \"kubernetes.io/projected/c4042633-8e4a-42e6-b311-dde8ef48716e-kube-api-access-f8q6p\") pod \"crc-debug-6sqtv\" (UID: \"c4042633-8e4a-42e6-b311-dde8ef48716e\") " pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.588450 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4042633-8e4a-42e6-b311-dde8ef48716e-host\") pod \"crc-debug-6sqtv\" (UID: \"c4042633-8e4a-42e6-b311-dde8ef48716e\") " pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.690990 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8q6p\" (UniqueName: \"kubernetes.io/projected/c4042633-8e4a-42e6-b311-dde8ef48716e-kube-api-access-f8q6p\") pod \"crc-debug-6sqtv\" (UID: \"c4042633-8e4a-42e6-b311-dde8ef48716e\") " pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.691768 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4042633-8e4a-42e6-b311-dde8ef48716e-host\") pod \"crc-debug-6sqtv\" (UID: \"c4042633-8e4a-42e6-b311-dde8ef48716e\") " pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.692010 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4042633-8e4a-42e6-b311-dde8ef48716e-host\") pod \"crc-debug-6sqtv\" (UID: \"c4042633-8e4a-42e6-b311-dde8ef48716e\") " pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.717018 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8q6p\" (UniqueName: \"kubernetes.io/projected/c4042633-8e4a-42e6-b311-dde8ef48716e-kube-api-access-f8q6p\") pod \"crc-debug-6sqtv\" (UID: \"c4042633-8e4a-42e6-b311-dde8ef48716e\") " pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.728264 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:21 crc kubenswrapper[4923]: W1009 11:06:21.767736 4923 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4042633_8e4a_42e6_b311_dde8ef48716e.slice/crio-588f5eb68a162863f3b9845960d9325f048a246d58fc24a604b9ba768f51d1f6 WatchSource:0}: Error finding container 588f5eb68a162863f3b9845960d9325f048a246d58fc24a604b9ba768f51d1f6: Status 404 returned error can't find the container with id 588f5eb68a162863f3b9845960d9325f048a246d58fc24a604b9ba768f51d1f6 Oct 09 11:06:21 crc kubenswrapper[4923]: I1009 11:06:21.967813 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" event={"ID":"c4042633-8e4a-42e6-b311-dde8ef48716e","Type":"ContainerStarted","Data":"588f5eb68a162863f3b9845960d9325f048a246d58fc24a604b9ba768f51d1f6"} Oct 09 11:06:28 crc kubenswrapper[4923]: I1009 11:06:28.602375 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:06:28 crc kubenswrapper[4923]: E1009 11:06:28.603561 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:06:35 crc kubenswrapper[4923]: I1009 11:06:35.119731 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" event={"ID":"c4042633-8e4a-42e6-b311-dde8ef48716e","Type":"ContainerStarted","Data":"6b8793560f0e13d38ea0d72ae5fc61f95648bf5e1ad33258f24f05da44816b70"} Oct 09 11:06:35 crc kubenswrapper[4923]: I1009 11:06:35.151312 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" podStartSLOduration=1.422083265 podStartE2EDuration="14.151277829s" podCreationTimestamp="2025-10-09 11:06:21 +0000 UTC" firstStartedPulling="2025-10-09 11:06:21.770532201 +0000 UTC m=+3667.838713957" lastFinishedPulling="2025-10-09 11:06:34.499726765 +0000 UTC m=+3680.567908521" observedRunningTime="2025-10-09 11:06:35.14083218 +0000 UTC m=+3681.209013936" watchObservedRunningTime="2025-10-09 11:06:35.151277829 +0000 UTC m=+3681.219459575" Oct 09 11:06:40 crc kubenswrapper[4923]: I1009 11:06:40.604561 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:06:40 crc kubenswrapper[4923]: E1009 11:06:40.605586 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:06:51 crc kubenswrapper[4923]: I1009 11:06:51.302978 4923 generic.go:334] "Generic (PLEG): container finished" podID="c4042633-8e4a-42e6-b311-dde8ef48716e" containerID="6b8793560f0e13d38ea0d72ae5fc61f95648bf5e1ad33258f24f05da44816b70" exitCode=0 Oct 09 11:06:51 crc kubenswrapper[4923]: I1009 11:06:51.303047 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" event={"ID":"c4042633-8e4a-42e6-b311-dde8ef48716e","Type":"ContainerDied","Data":"6b8793560f0e13d38ea0d72ae5fc61f95648bf5e1ad33258f24f05da44816b70"} Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.451133 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.488579 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-p2kmf/crc-debug-6sqtv"] Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.501366 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-p2kmf/crc-debug-6sqtv"] Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.577643 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8q6p\" (UniqueName: \"kubernetes.io/projected/c4042633-8e4a-42e6-b311-dde8ef48716e-kube-api-access-f8q6p\") pod \"c4042633-8e4a-42e6-b311-dde8ef48716e\" (UID: \"c4042633-8e4a-42e6-b311-dde8ef48716e\") " Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.577851 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4042633-8e4a-42e6-b311-dde8ef48716e-host\") pod \"c4042633-8e4a-42e6-b311-dde8ef48716e\" (UID: \"c4042633-8e4a-42e6-b311-dde8ef48716e\") " Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.577948 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c4042633-8e4a-42e6-b311-dde8ef48716e-host" (OuterVolumeSpecName: "host") pod "c4042633-8e4a-42e6-b311-dde8ef48716e" (UID: "c4042633-8e4a-42e6-b311-dde8ef48716e"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.578568 4923 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c4042633-8e4a-42e6-b311-dde8ef48716e-host\") on node \"crc\" DevicePath \"\"" Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.586061 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4042633-8e4a-42e6-b311-dde8ef48716e-kube-api-access-f8q6p" (OuterVolumeSpecName: "kube-api-access-f8q6p") pod "c4042633-8e4a-42e6-b311-dde8ef48716e" (UID: "c4042633-8e4a-42e6-b311-dde8ef48716e"). InnerVolumeSpecName "kube-api-access-f8q6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.601844 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:06:52 crc kubenswrapper[4923]: E1009 11:06:52.602388 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.613584 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4042633-8e4a-42e6-b311-dde8ef48716e" path="/var/lib/kubelet/pods/c4042633-8e4a-42e6-b311-dde8ef48716e/volumes" Oct 09 11:06:52 crc kubenswrapper[4923]: I1009 11:06:52.681412 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8q6p\" (UniqueName: \"kubernetes.io/projected/c4042633-8e4a-42e6-b311-dde8ef48716e-kube-api-access-f8q6p\") on node \"crc\" DevicePath \"\"" Oct 09 11:06:53 crc kubenswrapper[4923]: I1009 11:06:53.335284 4923 scope.go:117] "RemoveContainer" containerID="6b8793560f0e13d38ea0d72ae5fc61f95648bf5e1ad33258f24f05da44816b70" Oct 09 11:06:53 crc kubenswrapper[4923]: I1009 11:06:53.335543 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/crc-debug-6sqtv" Oct 09 11:06:53 crc kubenswrapper[4923]: I1009 11:06:53.841490 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-p2kmf/crc-debug-stdl5"] Oct 09 11:06:53 crc kubenswrapper[4923]: E1009 11:06:53.842641 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4042633-8e4a-42e6-b311-dde8ef48716e" containerName="container-00" Oct 09 11:06:53 crc kubenswrapper[4923]: I1009 11:06:53.842657 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4042633-8e4a-42e6-b311-dde8ef48716e" containerName="container-00" Oct 09 11:06:53 crc kubenswrapper[4923]: I1009 11:06:53.842913 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4042633-8e4a-42e6-b311-dde8ef48716e" containerName="container-00" Oct 09 11:06:53 crc kubenswrapper[4923]: I1009 11:06:53.843737 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:53 crc kubenswrapper[4923]: I1009 11:06:53.909679 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-host\") pod \"crc-debug-stdl5\" (UID: \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\") " pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:53 crc kubenswrapper[4923]: I1009 11:06:53.910042 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fs2b\" (UniqueName: \"kubernetes.io/projected/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-kube-api-access-6fs2b\") pod \"crc-debug-stdl5\" (UID: \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\") " pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:54 crc kubenswrapper[4923]: I1009 11:06:54.011723 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fs2b\" (UniqueName: \"kubernetes.io/projected/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-kube-api-access-6fs2b\") pod \"crc-debug-stdl5\" (UID: \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\") " pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:54 crc kubenswrapper[4923]: I1009 11:06:54.011892 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-host\") pod \"crc-debug-stdl5\" (UID: \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\") " pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:54 crc kubenswrapper[4923]: I1009 11:06:54.012058 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-host\") pod \"crc-debug-stdl5\" (UID: \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\") " pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:54 crc kubenswrapper[4923]: I1009 11:06:54.032029 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fs2b\" (UniqueName: \"kubernetes.io/projected/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-kube-api-access-6fs2b\") pod \"crc-debug-stdl5\" (UID: \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\") " pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:54 crc kubenswrapper[4923]: I1009 11:06:54.167108 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:54 crc kubenswrapper[4923]: I1009 11:06:54.356298 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p2kmf/crc-debug-stdl5" event={"ID":"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d","Type":"ContainerStarted","Data":"e1901ac84aee6d0323a4d57bebe451c3882a21ff592dd02120d2e92eeb1ad287"} Oct 09 11:06:55 crc kubenswrapper[4923]: I1009 11:06:55.369713 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p2kmf/crc-debug-stdl5" event={"ID":"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d","Type":"ContainerDied","Data":"5a10e7b60e938dc80d9f80f0eef67d01786e2ac0b5dc20fd70501fb4d015c87c"} Oct 09 11:06:55 crc kubenswrapper[4923]: I1009 11:06:55.369667 4923 generic.go:334] "Generic (PLEG): container finished" podID="f476a9e3-be16-4b40-ad20-7abfe0bb8d6d" containerID="5a10e7b60e938dc80d9f80f0eef67d01786e2ac0b5dc20fd70501fb4d015c87c" exitCode=1 Oct 09 11:06:55 crc kubenswrapper[4923]: I1009 11:06:55.418993 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-p2kmf/crc-debug-stdl5"] Oct 09 11:06:55 crc kubenswrapper[4923]: I1009 11:06:55.430088 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-p2kmf/crc-debug-stdl5"] Oct 09 11:06:56 crc kubenswrapper[4923]: I1009 11:06:56.498240 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:56 crc kubenswrapper[4923]: I1009 11:06:56.579789 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fs2b\" (UniqueName: \"kubernetes.io/projected/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-kube-api-access-6fs2b\") pod \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\" (UID: \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\") " Oct 09 11:06:56 crc kubenswrapper[4923]: I1009 11:06:56.580141 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-host\") pod \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\" (UID: \"f476a9e3-be16-4b40-ad20-7abfe0bb8d6d\") " Oct 09 11:06:56 crc kubenswrapper[4923]: I1009 11:06:56.580277 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-host" (OuterVolumeSpecName: "host") pod "f476a9e3-be16-4b40-ad20-7abfe0bb8d6d" (UID: "f476a9e3-be16-4b40-ad20-7abfe0bb8d6d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 11:06:56 crc kubenswrapper[4923]: I1009 11:06:56.580627 4923 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-host\") on node \"crc\" DevicePath \"\"" Oct 09 11:06:56 crc kubenswrapper[4923]: I1009 11:06:56.592295 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-kube-api-access-6fs2b" (OuterVolumeSpecName: "kube-api-access-6fs2b") pod "f476a9e3-be16-4b40-ad20-7abfe0bb8d6d" (UID: "f476a9e3-be16-4b40-ad20-7abfe0bb8d6d"). InnerVolumeSpecName "kube-api-access-6fs2b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:06:56 crc kubenswrapper[4923]: I1009 11:06:56.617142 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f476a9e3-be16-4b40-ad20-7abfe0bb8d6d" path="/var/lib/kubelet/pods/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d/volumes" Oct 09 11:06:56 crc kubenswrapper[4923]: I1009 11:06:56.684082 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fs2b\" (UniqueName: \"kubernetes.io/projected/f476a9e3-be16-4b40-ad20-7abfe0bb8d6d-kube-api-access-6fs2b\") on node \"crc\" DevicePath \"\"" Oct 09 11:06:57 crc kubenswrapper[4923]: I1009 11:06:57.390336 4923 scope.go:117] "RemoveContainer" containerID="5a10e7b60e938dc80d9f80f0eef67d01786e2ac0b5dc20fd70501fb4d015c87c" Oct 09 11:06:57 crc kubenswrapper[4923]: I1009 11:06:57.390425 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/crc-debug-stdl5" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.162272 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6jp4w"] Oct 09 11:06:58 crc kubenswrapper[4923]: E1009 11:06:58.165244 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f476a9e3-be16-4b40-ad20-7abfe0bb8d6d" containerName="container-00" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.165269 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="f476a9e3-be16-4b40-ad20-7abfe0bb8d6d" containerName="container-00" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.165541 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="f476a9e3-be16-4b40-ad20-7abfe0bb8d6d" containerName="container-00" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.167521 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.207398 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6jp4w"] Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.241922 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-catalog-content\") pod \"certified-operators-6jp4w\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.242601 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8v2jq\" (UniqueName: \"kubernetes.io/projected/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-kube-api-access-8v2jq\") pod \"certified-operators-6jp4w\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.242962 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-utilities\") pod \"certified-operators-6jp4w\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.345949 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-catalog-content\") pod \"certified-operators-6jp4w\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.346419 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8v2jq\" (UniqueName: \"kubernetes.io/projected/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-kube-api-access-8v2jq\") pod \"certified-operators-6jp4w\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.346535 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-utilities\") pod \"certified-operators-6jp4w\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.346590 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-catalog-content\") pod \"certified-operators-6jp4w\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.347490 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-utilities\") pod \"certified-operators-6jp4w\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.371841 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8v2jq\" (UniqueName: \"kubernetes.io/projected/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-kube-api-access-8v2jq\") pod \"certified-operators-6jp4w\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.501232 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:06:58 crc kubenswrapper[4923]: I1009 11:06:58.894001 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6jp4w"] Oct 09 11:06:59 crc kubenswrapper[4923]: I1009 11:06:59.418656 4923 generic.go:334] "Generic (PLEG): container finished" podID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerID="1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde" exitCode=0 Oct 09 11:06:59 crc kubenswrapper[4923]: I1009 11:06:59.419115 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jp4w" event={"ID":"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc","Type":"ContainerDied","Data":"1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde"} Oct 09 11:06:59 crc kubenswrapper[4923]: I1009 11:06:59.419159 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jp4w" event={"ID":"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc","Type":"ContainerStarted","Data":"903235fda830bf866ec5843d7979c60b218f28ceb70d26a8b80b513eeb6f5af3"} Oct 09 11:07:00 crc kubenswrapper[4923]: I1009 11:07:00.430767 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jp4w" event={"ID":"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc","Type":"ContainerStarted","Data":"8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef"} Oct 09 11:07:02 crc kubenswrapper[4923]: I1009 11:07:02.451932 4923 generic.go:334] "Generic (PLEG): container finished" podID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerID="8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef" exitCode=0 Oct 09 11:07:02 crc kubenswrapper[4923]: I1009 11:07:02.452026 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jp4w" event={"ID":"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc","Type":"ContainerDied","Data":"8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef"} Oct 09 11:07:04 crc kubenswrapper[4923]: I1009 11:07:04.482920 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jp4w" event={"ID":"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc","Type":"ContainerStarted","Data":"c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c"} Oct 09 11:07:04 crc kubenswrapper[4923]: I1009 11:07:04.517771 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6jp4w" podStartSLOduration=2.65344794 podStartE2EDuration="6.517734463s" podCreationTimestamp="2025-10-09 11:06:58 +0000 UTC" firstStartedPulling="2025-10-09 11:06:59.421163674 +0000 UTC m=+3705.489345430" lastFinishedPulling="2025-10-09 11:07:03.285450197 +0000 UTC m=+3709.353631953" observedRunningTime="2025-10-09 11:07:04.503146159 +0000 UTC m=+3710.571327935" watchObservedRunningTime="2025-10-09 11:07:04.517734463 +0000 UTC m=+3710.585916219" Oct 09 11:07:05 crc kubenswrapper[4923]: I1009 11:07:05.602817 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:07:05 crc kubenswrapper[4923]: E1009 11:07:05.603575 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:07:08 crc kubenswrapper[4923]: I1009 11:07:08.501361 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:07:08 crc kubenswrapper[4923]: I1009 11:07:08.503527 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:07:08 crc kubenswrapper[4923]: I1009 11:07:08.568215 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:07:08 crc kubenswrapper[4923]: I1009 11:07:08.624099 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:07:08 crc kubenswrapper[4923]: I1009 11:07:08.808421 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6jp4w"] Oct 09 11:07:10 crc kubenswrapper[4923]: I1009 11:07:10.538674 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6jp4w" podUID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerName="registry-server" containerID="cri-o://c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c" gracePeriod=2 Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.059165 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.162895 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8v2jq\" (UniqueName: \"kubernetes.io/projected/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-kube-api-access-8v2jq\") pod \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.163147 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-utilities\") pod \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.163222 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-catalog-content\") pod \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\" (UID: \"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc\") " Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.164343 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-utilities" (OuterVolumeSpecName: "utilities") pod "b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" (UID: "b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.174334 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-kube-api-access-8v2jq" (OuterVolumeSpecName: "kube-api-access-8v2jq") pod "b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" (UID: "b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc"). InnerVolumeSpecName "kube-api-access-8v2jq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.214669 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" (UID: "b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.267196 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.267235 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8v2jq\" (UniqueName: \"kubernetes.io/projected/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-kube-api-access-8v2jq\") on node \"crc\" DevicePath \"\"" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.267248 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.550254 4923 generic.go:334] "Generic (PLEG): container finished" podID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerID="c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c" exitCode=0 Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.550312 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jp4w" event={"ID":"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc","Type":"ContainerDied","Data":"c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c"} Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.550353 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6jp4w" event={"ID":"b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc","Type":"ContainerDied","Data":"903235fda830bf866ec5843d7979c60b218f28ceb70d26a8b80b513eeb6f5af3"} Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.550406 4923 scope.go:117] "RemoveContainer" containerID="c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.550597 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6jp4w" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.591272 4923 scope.go:117] "RemoveContainer" containerID="8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.596188 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6jp4w"] Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.614125 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6jp4w"] Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.624121 4923 scope.go:117] "RemoveContainer" containerID="1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.692623 4923 scope.go:117] "RemoveContainer" containerID="c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c" Oct 09 11:07:11 crc kubenswrapper[4923]: E1009 11:07:11.693137 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c\": container with ID starting with c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c not found: ID does not exist" containerID="c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.693171 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c"} err="failed to get container status \"c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c\": rpc error: code = NotFound desc = could not find container \"c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c\": container with ID starting with c9dad4bb77a587289fe757fbe1abd7a77b176b5cf76306bfa510e6c0c636c67c not found: ID does not exist" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.693200 4923 scope.go:117] "RemoveContainer" containerID="8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef" Oct 09 11:07:11 crc kubenswrapper[4923]: E1009 11:07:11.693661 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef\": container with ID starting with 8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef not found: ID does not exist" containerID="8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.693683 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef"} err="failed to get container status \"8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef\": rpc error: code = NotFound desc = could not find container \"8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef\": container with ID starting with 8ab256c9df1d7ede0cd7aaef37de5128c97a5e758bcac96f7b59787fe348f7ef not found: ID does not exist" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.693703 4923 scope.go:117] "RemoveContainer" containerID="1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde" Oct 09 11:07:11 crc kubenswrapper[4923]: E1009 11:07:11.693991 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde\": container with ID starting with 1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde not found: ID does not exist" containerID="1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde" Oct 09 11:07:11 crc kubenswrapper[4923]: I1009 11:07:11.694014 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde"} err="failed to get container status \"1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde\": rpc error: code = NotFound desc = could not find container \"1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde\": container with ID starting with 1d486fe8576cc8bb094324a157b35ce8eb3f0c8436ca2390b0260a2210579cde not found: ID does not exist" Oct 09 11:07:12 crc kubenswrapper[4923]: I1009 11:07:12.612913 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" path="/var/lib/kubelet/pods/b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc/volumes" Oct 09 11:07:18 crc kubenswrapper[4923]: I1009 11:07:18.602208 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:07:18 crc kubenswrapper[4923]: E1009 11:07:18.603224 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:07:23 crc kubenswrapper[4923]: I1009 11:07:23.908459 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9f849b6cd-m82rj_45877ffa-4abb-489a-9bad-770cbb5f231a/barbican-api/0.log" Oct 09 11:07:23 crc kubenswrapper[4923]: I1009 11:07:23.912645 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-9f849b6cd-m82rj_45877ffa-4abb-489a-9bad-770cbb5f231a/barbican-api-log/0.log" Oct 09 11:07:24 crc kubenswrapper[4923]: I1009 11:07:24.092708 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7bc74dbf64-gj94f_8334ffed-82c7-483f-a947-6f3c132965f8/barbican-keystone-listener/0.log" Oct 09 11:07:24 crc kubenswrapper[4923]: I1009 11:07:24.175217 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7bc74dbf64-gj94f_8334ffed-82c7-483f-a947-6f3c132965f8/barbican-keystone-listener-log/0.log" Oct 09 11:07:24 crc kubenswrapper[4923]: I1009 11:07:24.349075 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-55d7d7d987-nr6w5_169c7b93-c284-4cf2-bc5c-c5f3a7c7277d/barbican-worker/0.log" Oct 09 11:07:24 crc kubenswrapper[4923]: I1009 11:07:24.379278 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-55d7d7d987-nr6w5_169c7b93-c284-4cf2-bc5c-c5f3a7c7277d/barbican-worker-log/0.log" Oct 09 11:07:24 crc kubenswrapper[4923]: I1009 11:07:24.584184 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-ssdb6_3f45569e-63f8-4693-9b2c-6b6e8d75067b/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:24 crc kubenswrapper[4923]: I1009 11:07:24.827571 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_dba22ce5-7738-4868-b2f2-d52f43bccbe5/ceilometer-central-agent/0.log" Oct 09 11:07:24 crc kubenswrapper[4923]: I1009 11:07:24.831266 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_dba22ce5-7738-4868-b2f2-d52f43bccbe5/ceilometer-notification-agent/0.log" Oct 09 11:07:24 crc kubenswrapper[4923]: I1009 11:07:24.889428 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_dba22ce5-7738-4868-b2f2-d52f43bccbe5/proxy-httpd/0.log" Oct 09 11:07:25 crc kubenswrapper[4923]: I1009 11:07:25.006321 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_dba22ce5-7738-4868-b2f2-d52f43bccbe5/sg-core/0.log" Oct 09 11:07:25 crc kubenswrapper[4923]: I1009 11:07:25.119708 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-v6cm7_43c42b69-73b7-4758-92cc-431d2837fb39/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:25 crc kubenswrapper[4923]: I1009 11:07:25.333735 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-62zfk_18edca22-01b1-4d52-ac43-d93611eb5464/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:25 crc kubenswrapper[4923]: I1009 11:07:25.541523 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62/cinder-api-log/0.log" Oct 09 11:07:25 crc kubenswrapper[4923]: I1009 11:07:25.543472 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_07de3fcd-6ca3-4c5a-a57b-97a51ffe6d62/cinder-api/0.log" Oct 09 11:07:25 crc kubenswrapper[4923]: I1009 11:07:25.752710 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5d650442-1c4b-48cd-8b94-e9dd6e51cae3/probe/0.log" Oct 09 11:07:25 crc kubenswrapper[4923]: I1009 11:07:25.912662 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_5d650442-1c4b-48cd-8b94-e9dd6e51cae3/cinder-backup/0.log" Oct 09 11:07:26 crc kubenswrapper[4923]: I1009 11:07:26.067176 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e99cc2d5-9514-4cd1-9fb7-708774861057/cinder-scheduler/0.log" Oct 09 11:07:26 crc kubenswrapper[4923]: I1009 11:07:26.199154 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e99cc2d5-9514-4cd1-9fb7-708774861057/probe/0.log" Oct 09 11:07:26 crc kubenswrapper[4923]: I1009 11:07:26.344056 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_23f48866-c443-458e-9a5a-dd2243b86249/cinder-volume/0.log" Oct 09 11:07:26 crc kubenswrapper[4923]: I1009 11:07:26.423554 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_23f48866-c443-458e-9a5a-dd2243b86249/probe/0.log" Oct 09 11:07:26 crc kubenswrapper[4923]: I1009 11:07:26.870484 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-kzf7w_d8452ab6-f295-464d-bf20-c4a0d7f6ce64/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:27 crc kubenswrapper[4923]: I1009 11:07:27.028876 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-gjw68_c66ce1f0-4c37-4e5d-ae58-bbebdfa66925/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:27 crc kubenswrapper[4923]: I1009 11:07:27.272388 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55d8975557-gm28r_0cccdeaa-06bb-465b-8354-79074c54ada9/init/0.log" Oct 09 11:07:27 crc kubenswrapper[4923]: I1009 11:07:27.496876 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55d8975557-gm28r_0cccdeaa-06bb-465b-8354-79074c54ada9/dnsmasq-dns/0.log" Oct 09 11:07:27 crc kubenswrapper[4923]: I1009 11:07:27.528374 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-55d8975557-gm28r_0cccdeaa-06bb-465b-8354-79074c54ada9/init/0.log" Oct 09 11:07:27 crc kubenswrapper[4923]: I1009 11:07:27.534512 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5a6db3b3-10aa-4af5-958a-15f9ca7f5102/glance-httpd/0.log" Oct 09 11:07:27 crc kubenswrapper[4923]: I1009 11:07:27.739388 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_5a6db3b3-10aa-4af5-958a-15f9ca7f5102/glance-log/0.log" Oct 09 11:07:27 crc kubenswrapper[4923]: I1009 11:07:27.796690 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5/glance-httpd/0.log" Oct 09 11:07:27 crc kubenswrapper[4923]: I1009 11:07:27.940397 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_9d60ad38-2e8a-4cd9-82ab-98bc16ac9fb5/glance-log/0.log" Oct 09 11:07:28 crc kubenswrapper[4923]: I1009 11:07:28.187708 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5dd576c454-s7hpl_4d75434e-89d3-464b-af06-fb0ac9f92fd8/horizon/0.log" Oct 09 11:07:28 crc kubenswrapper[4923]: I1009 11:07:28.391204 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-dzcw5_b8e09b1b-4163-4570-96c6-aca084f2ea50/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:28 crc kubenswrapper[4923]: I1009 11:07:28.468931 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5dd576c454-s7hpl_4d75434e-89d3-464b-af06-fb0ac9f92fd8/horizon-log/0.log" Oct 09 11:07:28 crc kubenswrapper[4923]: I1009 11:07:28.584581 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-fk6lc_ec0a7dfb-a4f5-41a4-a83c-d3d3eda6db1f/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:28 crc kubenswrapper[4923]: I1009 11:07:28.805452 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c7cc6d588-zpbfk_4a4c4530-4816-41ac-bdfc-945bb60600c7/keystone-api/0.log" Oct 09 11:07:28 crc kubenswrapper[4923]: I1009 11:07:28.817547 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29333461-8cpkk_2917616d-7104-4c9e-9fa2-2590c69866b2/keystone-cron/0.log" Oct 09 11:07:29 crc kubenswrapper[4923]: I1009 11:07:29.072687 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-2f26f_81549d8a-1bfc-4261-b7bc-7edb5b2d9a45/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:29 crc kubenswrapper[4923]: I1009 11:07:29.078613 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_73d694c1-4a65-457d-8033-8b829df1ccbb/kube-state-metrics/0.log" Oct 09 11:07:29 crc kubenswrapper[4923]: I1009 11:07:29.461292 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-37f2-account-create-64wqp_53939d29-b8dc-4a59-a7b7-37a8675ced9a/mariadb-account-create/0.log" Oct 09 11:07:29 crc kubenswrapper[4923]: I1009 11:07:29.724344 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_0afc3325-d376-42d0-a3b2-650f4f707d56/manila-api-log/0.log" Oct 09 11:07:29 crc kubenswrapper[4923]: I1009 11:07:29.731348 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_0afc3325-d376-42d0-a3b2-650f4f707d56/manila-api/0.log" Oct 09 11:07:29 crc kubenswrapper[4923]: I1009 11:07:29.824351 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-db-create-mxv9k_f093ece7-e8f5-4beb-8f89-612afece0428/mariadb-database-create/0.log" Oct 09 11:07:29 crc kubenswrapper[4923]: I1009 11:07:29.933368 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-db-sync-z6tpv_e665c786-66db-460a-9bd8-c1ec9ecfb978/manila-db-sync/0.log" Oct 09 11:07:30 crc kubenswrapper[4923]: I1009 11:07:30.136918 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_38e3cad2-d7c0-463e-a63c-a197183d870b/manila-scheduler/0.log" Oct 09 11:07:30 crc kubenswrapper[4923]: I1009 11:07:30.179056 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_38e3cad2-d7c0-463e-a63c-a197183d870b/probe/0.log" Oct 09 11:07:30 crc kubenswrapper[4923]: I1009 11:07:30.349843 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7fb17c25-9276-4577-b9c1-56ca34a4dece/manila-share/0.log" Oct 09 11:07:30 crc kubenswrapper[4923]: I1009 11:07:30.412897 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7fb17c25-9276-4577-b9c1-56ca34a4dece/probe/0.log" Oct 09 11:07:30 crc kubenswrapper[4923]: I1009 11:07:30.739334 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5fbfdd98cc-7gf4c_ac145f4f-e2db-4f2f-8842-7256514a4ee2/neutron-api/0.log" Oct 09 11:07:30 crc kubenswrapper[4923]: I1009 11:07:30.760661 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5fbfdd98cc-7gf4c_ac145f4f-e2db-4f2f-8842-7256514a4ee2/neutron-httpd/0.log" Oct 09 11:07:30 crc kubenswrapper[4923]: I1009 11:07:30.980564 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-hdswp_480a61a2-d67f-45ed-b2ef-d24b8b750035/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:31 crc kubenswrapper[4923]: I1009 11:07:31.437583 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_555f5900-a091-41e5-982e-6d8e46d1fbe5/nova-api-log/0.log" Oct 09 11:07:31 crc kubenswrapper[4923]: I1009 11:07:31.602030 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:07:31 crc kubenswrapper[4923]: E1009 11:07:31.602384 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:07:31 crc kubenswrapper[4923]: I1009 11:07:31.709571 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_555f5900-a091-41e5-982e-6d8e46d1fbe5/nova-api-api/0.log" Oct 09 11:07:31 crc kubenswrapper[4923]: I1009 11:07:31.769818 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_3f52a73f-7675-4cd5-ae86-77bd49333940/nova-cell0-conductor-conductor/0.log" Oct 09 11:07:32 crc kubenswrapper[4923]: I1009 11:07:32.173701 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_99c043f4-dce1-4779-8b34-6b988207c784/nova-cell1-conductor-conductor/0.log" Oct 09 11:07:32 crc kubenswrapper[4923]: I1009 11:07:32.217594 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_8b60a796-c363-488e-9f05-1dedbf605ab5/nova-cell1-novncproxy-novncproxy/0.log" Oct 09 11:07:32 crc kubenswrapper[4923]: I1009 11:07:32.533832 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-6mr8h_c2ee9657-939a-4acf-a25b-eb315ebec083/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:32 crc kubenswrapper[4923]: I1009 11:07:32.631129 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e5e67cba-1bce-4d4a-b1f0-079ebd765787/nova-metadata-log/0.log" Oct 09 11:07:33 crc kubenswrapper[4923]: I1009 11:07:33.055451 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_37f5090b-66d4-409a-b7d1-013eda9b7f45/nova-scheduler-scheduler/0.log" Oct 09 11:07:33 crc kubenswrapper[4923]: I1009 11:07:33.516966 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4d9b867d-d4b6-4fe5-ad0b-4b5909577d25/mysql-bootstrap/0.log" Oct 09 11:07:33 crc kubenswrapper[4923]: I1009 11:07:33.748644 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4d9b867d-d4b6-4fe5-ad0b-4b5909577d25/galera/0.log" Oct 09 11:07:33 crc kubenswrapper[4923]: I1009 11:07:33.774387 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_4d9b867d-d4b6-4fe5-ad0b-4b5909577d25/mysql-bootstrap/0.log" Oct 09 11:07:34 crc kubenswrapper[4923]: I1009 11:07:34.086992 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d/mysql-bootstrap/0.log" Oct 09 11:07:34 crc kubenswrapper[4923]: I1009 11:07:34.309997 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d/mysql-bootstrap/0.log" Oct 09 11:07:34 crc kubenswrapper[4923]: I1009 11:07:34.369660 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_3ecb3847-ac13-4e9a-b8db-abafa3ca6a0d/galera/0.log" Oct 09 11:07:34 crc kubenswrapper[4923]: I1009 11:07:34.372413 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e5e67cba-1bce-4d4a-b1f0-079ebd765787/nova-metadata-metadata/0.log" Oct 09 11:07:34 crc kubenswrapper[4923]: I1009 11:07:34.693527 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_bb06dc17-1098-420f-9106-4ea25ae8bd9d/openstackclient/0.log" Oct 09 11:07:34 crc kubenswrapper[4923]: I1009 11:07:34.851149 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-jgs4j_4c323704-d81e-4c9b-a58d-b9943c2f55d7/ovn-controller/0.log" Oct 09 11:07:35 crc kubenswrapper[4923]: I1009 11:07:35.001308 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-t8h9w_ce6ae725-8786-4973-bafe-8d6548ee51bb/openstack-network-exporter/0.log" Oct 09 11:07:35 crc kubenswrapper[4923]: I1009 11:07:35.331562 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m429c_1d1e3e9d-552f-441c-ae67-8af6e4fda12b/ovsdb-server-init/0.log" Oct 09 11:07:35 crc kubenswrapper[4923]: I1009 11:07:35.510975 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m429c_1d1e3e9d-552f-441c-ae67-8af6e4fda12b/ovsdb-server-init/0.log" Oct 09 11:07:35 crc kubenswrapper[4923]: I1009 11:07:35.591215 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m429c_1d1e3e9d-552f-441c-ae67-8af6e4fda12b/ovs-vswitchd/0.log" Oct 09 11:07:35 crc kubenswrapper[4923]: I1009 11:07:35.602133 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-m429c_1d1e3e9d-552f-441c-ae67-8af6e4fda12b/ovsdb-server/0.log" Oct 09 11:07:35 crc kubenswrapper[4923]: I1009 11:07:35.945257 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-z8dvq_eedf18b1-d421-493d-838c-965b961d21a3/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:36 crc kubenswrapper[4923]: I1009 11:07:36.083476 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9ef458c9-e8d0-4251-924b-c46ef4cb203b/openstack-network-exporter/0.log" Oct 09 11:07:36 crc kubenswrapper[4923]: I1009 11:07:36.092013 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9ef458c9-e8d0-4251-924b-c46ef4cb203b/ovn-northd/0.log" Oct 09 11:07:36 crc kubenswrapper[4923]: I1009 11:07:36.253318 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_da46f3c8-e318-4f39-94e8-814eb2d775bc/openstack-network-exporter/0.log" Oct 09 11:07:36 crc kubenswrapper[4923]: I1009 11:07:36.383195 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_da46f3c8-e318-4f39-94e8-814eb2d775bc/ovsdbserver-nb/0.log" Oct 09 11:07:36 crc kubenswrapper[4923]: I1009 11:07:36.522431 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_46f24e5a-a778-40c2-837e-e5cfda1992ff/openstack-network-exporter/0.log" Oct 09 11:07:36 crc kubenswrapper[4923]: I1009 11:07:36.543734 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_46f24e5a-a778-40c2-837e-e5cfda1992ff/ovsdbserver-sb/0.log" Oct 09 11:07:36 crc kubenswrapper[4923]: I1009 11:07:36.740071 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-84bc56c5fd-mzbzg_c8c099f6-0c20-40d9-a47e-35150638b521/placement-api/0.log" Oct 09 11:07:36 crc kubenswrapper[4923]: I1009 11:07:36.891078 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-84bc56c5fd-mzbzg_c8c099f6-0c20-40d9-a47e-35150638b521/placement-log/0.log" Oct 09 11:07:37 crc kubenswrapper[4923]: I1009 11:07:37.021008 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b1b005e0-1c67-42a3-97aa-3222f4218dca/setup-container/0.log" Oct 09 11:07:37 crc kubenswrapper[4923]: I1009 11:07:37.568022 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b1b005e0-1c67-42a3-97aa-3222f4218dca/setup-container/0.log" Oct 09 11:07:37 crc kubenswrapper[4923]: I1009 11:07:37.649310 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_b1b005e0-1c67-42a3-97aa-3222f4218dca/rabbitmq/0.log" Oct 09 11:07:37 crc kubenswrapper[4923]: I1009 11:07:37.674926 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8dbce0ef-9426-4065-9749-633e7a5e4248/setup-container/0.log" Oct 09 11:07:37 crc kubenswrapper[4923]: I1009 11:07:37.972619 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-zsxx6_6494d91a-6ead-46aa-b724-20fbca0b5319/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:37 crc kubenswrapper[4923]: I1009 11:07:37.991337 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8dbce0ef-9426-4065-9749-633e7a5e4248/setup-container/0.log" Oct 09 11:07:38 crc kubenswrapper[4923]: I1009 11:07:38.001490 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_8dbce0ef-9426-4065-9749-633e7a5e4248/rabbitmq/0.log" Oct 09 11:07:38 crc kubenswrapper[4923]: I1009 11:07:38.235167 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-ctf88_9921186f-446f-485d-b39b-e86be6df0511/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:38 crc kubenswrapper[4923]: I1009 11:07:38.429179 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-wjkcc_6881fbdc-1adc-4922-aca4-484a5fa2a348/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:38 crc kubenswrapper[4923]: I1009 11:07:38.663242 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-c28nb_7fc44ca7-2ff1-4677-b3d1-860332fef1c4/ssh-known-hosts-edpm-deployment/0.log" Oct 09 11:07:38 crc kubenswrapper[4923]: I1009 11:07:38.782513 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-sncpq_b1ab7440-9c9d-4ba4-92c3-fe19bccbde63/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 11:07:42 crc kubenswrapper[4923]: I1009 11:07:42.616287 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:07:42 crc kubenswrapper[4923]: E1009 11:07:42.621279 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:07:42 crc kubenswrapper[4923]: I1009 11:07:42.754561 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c68f8faa-8010-49a5-949a-94933174149b/memcached/0.log" Oct 09 11:07:53 crc kubenswrapper[4923]: I1009 11:07:53.601359 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:07:53 crc kubenswrapper[4923]: E1009 11:07:53.602248 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.000504 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp_1a2f9a94-3b55-48ee-aa73-d22e8bde118c/util/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.212252 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp_1a2f9a94-3b55-48ee-aa73-d22e8bde118c/util/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.251166 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp_1a2f9a94-3b55-48ee-aa73-d22e8bde118c/pull/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.279774 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp_1a2f9a94-3b55-48ee-aa73-d22e8bde118c/pull/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.460034 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp_1a2f9a94-3b55-48ee-aa73-d22e8bde118c/pull/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.460811 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp_1a2f9a94-3b55-48ee-aa73-d22e8bde118c/util/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.483451 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_184bca519f21fd0fd55ec08aa4c93472bde537c6b6fd14be86cecc7c99msbdp_1a2f9a94-3b55-48ee-aa73-d22e8bde118c/extract/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.602150 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:08:08 crc kubenswrapper[4923]: E1009 11:08:08.602450 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.675668 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-h294t_91f871ca-9617-49c0-877f-ee0bcf938bee/kube-rbac-proxy/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.751521 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lpvgj_4466effc-4c4f-4023-94bc-d1e2f48d5d3b/kube-rbac-proxy/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.759711 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-658bdf4b74-h294t_91f871ca-9617-49c0-877f-ee0bcf938bee/manager/0.log" Oct 09 11:08:08 crc kubenswrapper[4923]: I1009 11:08:08.984596 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-7b7fb68549-lpvgj_4466effc-4c4f-4023-94bc-d1e2f48d5d3b/manager/0.log" Oct 09 11:08:09 crc kubenswrapper[4923]: I1009 11:08:09.015437 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-lcs47_4609ba20-418f-4134-b671-10154112f387/manager/0.log" Oct 09 11:08:09 crc kubenswrapper[4923]: I1009 11:08:09.017595 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-85d5d9dd78-lcs47_4609ba20-418f-4134-b671-10154112f387/kube-rbac-proxy/0.log" Oct 09 11:08:09 crc kubenswrapper[4923]: I1009 11:08:09.271517 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-fhsk4_1438c7f5-06a5-49d0-be96-7fbee10b3010/kube-rbac-proxy/0.log" Oct 09 11:08:09 crc kubenswrapper[4923]: I1009 11:08:09.281738 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-84b9b84486-fhsk4_1438c7f5-06a5-49d0-be96-7fbee10b3010/manager/0.log" Oct 09 11:08:09 crc kubenswrapper[4923]: I1009 11:08:09.436119 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-zjhgq_36c97947-5f3f-4d0e-8125-0663f04f9eaf/kube-rbac-proxy/0.log" Oct 09 11:08:09 crc kubenswrapper[4923]: I1009 11:08:09.490502 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-858f76bbdd-zjhgq_36c97947-5f3f-4d0e-8125-0663f04f9eaf/manager/0.log" Oct 09 11:08:09 crc kubenswrapper[4923]: I1009 11:08:09.546109 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-gdprs_a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8/kube-rbac-proxy/0.log" Oct 09 11:08:09 crc kubenswrapper[4923]: I1009 11:08:09.648468 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-7ffbcb7588-gdprs_a4de9e67-0c5c-4c1e-8eba-7e0d14599dc8/manager/0.log" Oct 09 11:08:10 crc kubenswrapper[4923]: I1009 11:08:10.306215 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-td8lq_c5fba54a-5e97-419a-8950-65bd3908aadb/kube-rbac-proxy/0.log" Oct 09 11:08:10 crc kubenswrapper[4923]: I1009 11:08:10.510392 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-656bcbd775-td8lq_c5fba54a-5e97-419a-8950-65bd3908aadb/manager/0.log" Oct 09 11:08:10 crc kubenswrapper[4923]: I1009 11:08:10.554885 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-n8tdz_c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72/kube-rbac-proxy/0.log" Oct 09 11:08:10 crc kubenswrapper[4923]: I1009 11:08:10.578236 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-9c5c78d49-n8tdz_c7ebbe8d-b85b-479c-9fd8-e7ad2197bc72/manager/0.log" Oct 09 11:08:10 crc kubenswrapper[4923]: I1009 11:08:10.750871 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-qrh4j_2ad61b43-c2ba-4dc9-8376-e75817b9e153/kube-rbac-proxy/0.log" Oct 09 11:08:10 crc kubenswrapper[4923]: I1009 11:08:10.846845 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-55b6b7c7b8-qrh4j_2ad61b43-c2ba-4dc9-8376-e75817b9e153/manager/0.log" Oct 09 11:08:10 crc kubenswrapper[4923]: I1009 11:08:10.938397 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-jz6s9_3fd48b51-1de2-4d9f-a790-55c95638fc2f/kube-rbac-proxy/0.log" Oct 09 11:08:11 crc kubenswrapper[4923]: I1009 11:08:11.053071 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5f67fbc655-jz6s9_3fd48b51-1de2-4d9f-a790-55c95638fc2f/manager/0.log" Oct 09 11:08:11 crc kubenswrapper[4923]: I1009 11:08:11.151532 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-md5f2_575ef7c8-b9ea-47ae-aa33-8c87596fff08/kube-rbac-proxy/0.log" Oct 09 11:08:11 crc kubenswrapper[4923]: I1009 11:08:11.214966 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-f9fb45f8f-md5f2_575ef7c8-b9ea-47ae-aa33-8c87596fff08/manager/0.log" Oct 09 11:08:11 crc kubenswrapper[4923]: I1009 11:08:11.293828 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-tp6xl_a44d21c5-6bf2-445e-9d72-2463571f1cf3/kube-rbac-proxy/0.log" Oct 09 11:08:11 crc kubenswrapper[4923]: I1009 11:08:11.466687 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-79d585cb66-tp6xl_a44d21c5-6bf2-445e-9d72-2463571f1cf3/manager/0.log" Oct 09 11:08:11 crc kubenswrapper[4923]: I1009 11:08:11.561485 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-55fjr_42a720da-4703-4d8b-9046-6c8e1643c880/kube-rbac-proxy/0.log" Oct 09 11:08:11 crc kubenswrapper[4923]: I1009 11:08:11.682952 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5df598886f-55fjr_42a720da-4703-4d8b-9046-6c8e1643c880/manager/0.log" Oct 09 11:08:11 crc kubenswrapper[4923]: I1009 11:08:11.768255 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-pbjk9_120ce970-aff3-40cb-8724-957a034fb914/kube-rbac-proxy/0.log" Oct 09 11:08:11 crc kubenswrapper[4923]: I1009 11:08:11.814142 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-69fdcfc5f5-pbjk9_120ce970-aff3-40cb-8724-957a034fb914/manager/0.log" Oct 09 11:08:12 crc kubenswrapper[4923]: I1009 11:08:12.476319 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-677c5f5bffmcv74_29e93b24-6e22-4ade-9fd9-1742186d21fd/kube-rbac-proxy/0.log" Oct 09 11:08:12 crc kubenswrapper[4923]: I1009 11:08:12.508086 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-677c5f5bffmcv74_29e93b24-6e22-4ade-9fd9-1742186d21fd/manager/0.log" Oct 09 11:08:12 crc kubenswrapper[4923]: I1009 11:08:12.518707 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7d6957655c-h7b88_2ac22a8a-f293-4c68-85be-286bc263e77b/kube-rbac-proxy/0.log" Oct 09 11:08:12 crc kubenswrapper[4923]: I1009 11:08:12.845469 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5c7bf7948-jnnfj_b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0/kube-rbac-proxy/0.log" Oct 09 11:08:12 crc kubenswrapper[4923]: I1009 11:08:12.845849 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5c7bf7948-jnnfj_b88cbcc7-3c1f-4228-8563-ee01cd9cf5f0/operator/0.log" Oct 09 11:08:13 crc kubenswrapper[4923]: I1009 11:08:13.062144 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-bbtq2_8ba088a5-a6d3-48ab-8966-efd7be3f7981/registry-server/0.log" Oct 09 11:08:13 crc kubenswrapper[4923]: I1009 11:08:13.100512 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-f89mg_6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed/kube-rbac-proxy/0.log" Oct 09 11:08:13 crc kubenswrapper[4923]: I1009 11:08:13.244373 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-79db49b9fb-f89mg_6d7ad3b7-04a4-41d0-90fe-15f32fa9c0ed/manager/0.log" Oct 09 11:08:13 crc kubenswrapper[4923]: I1009 11:08:13.353277 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-2cvzn_ada28834-1248-46cb-b482-4a2d8ee77520/kube-rbac-proxy/0.log" Oct 09 11:08:13 crc kubenswrapper[4923]: I1009 11:08:13.526341 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-68b6c87b68-2cvzn_ada28834-1248-46cb-b482-4a2d8ee77520/manager/0.log" Oct 09 11:08:13 crc kubenswrapper[4923]: I1009 11:08:13.710960 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-ktz5t_cd75b71e-8663-4e84-a99f-2d902135917d/operator/0.log" Oct 09 11:08:13 crc kubenswrapper[4923]: I1009 11:08:13.786676 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-xzpv6_c5e2cc1d-7b01-4316-a960-5c47a420d6fd/kube-rbac-proxy/0.log" Oct 09 11:08:13 crc kubenswrapper[4923]: I1009 11:08:13.855071 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-db6d7f97b-xzpv6_c5e2cc1d-7b01-4316-a960-5c47a420d6fd/manager/0.log" Oct 09 11:08:13 crc kubenswrapper[4923]: I1009 11:08:13.963990 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-hw8r6_9e8c5a78-2a63-4219-8293-ea6d866a2d1b/kube-rbac-proxy/0.log" Oct 09 11:08:14 crc kubenswrapper[4923]: I1009 11:08:14.086040 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-7d6957655c-h7b88_2ac22a8a-f293-4c68-85be-286bc263e77b/manager/0.log" Oct 09 11:08:14 crc kubenswrapper[4923]: I1009 11:08:14.248021 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-67cfc6749b-hw8r6_9e8c5a78-2a63-4219-8293-ea6d866a2d1b/manager/0.log" Oct 09 11:08:14 crc kubenswrapper[4923]: I1009 11:08:14.262149 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-776875b78c-5975g_c703820f-a08c-4666-abc7-48954e40d927/kube-rbac-proxy/0.log" Oct 09 11:08:14 crc kubenswrapper[4923]: I1009 11:08:14.272274 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-776875b78c-5975g_c703820f-a08c-4666-abc7-48954e40d927/manager/1.log" Oct 09 11:08:14 crc kubenswrapper[4923]: I1009 11:08:14.401198 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-776875b78c-5975g_c703820f-a08c-4666-abc7-48954e40d927/manager/0.log" Oct 09 11:08:14 crc kubenswrapper[4923]: I1009 11:08:14.487523 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-54cbp_c0d0a77b-c472-4690-b169-becf1ea318dc/kube-rbac-proxy/0.log" Oct 09 11:08:14 crc kubenswrapper[4923]: I1009 11:08:14.498729 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-7f554bff7b-54cbp_c0d0a77b-c472-4690-b169-becf1ea318dc/manager/0.log" Oct 09 11:08:21 crc kubenswrapper[4923]: I1009 11:08:21.602063 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:08:21 crc kubenswrapper[4923]: E1009 11:08:21.602908 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:08:32 crc kubenswrapper[4923]: I1009 11:08:32.602469 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:08:32 crc kubenswrapper[4923]: E1009 11:08:32.603666 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:08:32 crc kubenswrapper[4923]: I1009 11:08:32.909929 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-z97pw_c37f83cf-f85a-4872-8e9e-ed37253c753f/control-plane-machine-set-operator/0.log" Oct 09 11:08:33 crc kubenswrapper[4923]: I1009 11:08:33.081657 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9sd7w_d06b0170-f718-4285-a25b-b5c07868e3c2/machine-api-operator/0.log" Oct 09 11:08:33 crc kubenswrapper[4923]: I1009 11:08:33.108861 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9sd7w_d06b0170-f718-4285-a25b-b5c07868e3c2/kube-rbac-proxy/0.log" Oct 09 11:08:45 crc kubenswrapper[4923]: I1009 11:08:45.656593 4923 generic.go:334] "Generic (PLEG): container finished" podID="c703820f-a08c-4666-abc7-48954e40d927" containerID="670db4a03e2a459f6e766d46d957174f0b06617bc7aeac6d8ff72e27c423a1f6" exitCode=1 Oct 09 11:08:45 crc kubenswrapper[4923]: I1009 11:08:45.656677 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerDied","Data":"670db4a03e2a459f6e766d46d957174f0b06617bc7aeac6d8ff72e27c423a1f6"} Oct 09 11:08:45 crc kubenswrapper[4923]: I1009 11:08:45.657515 4923 scope.go:117] "RemoveContainer" containerID="9db60a6471e24ee9345cd8dfc83b94aa0a95eff63f1d886e992291d3e879ea8a" Oct 09 11:08:45 crc kubenswrapper[4923]: I1009 11:08:45.658596 4923 scope.go:117] "RemoveContainer" containerID="670db4a03e2a459f6e766d46d957174f0b06617bc7aeac6d8ff72e27c423a1f6" Oct 09 11:08:45 crc kubenswrapper[4923]: E1009 11:08:45.659361 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=test-operator-controller-manager-776875b78c-5975g_openstack-operators(c703820f-a08c-4666-abc7-48954e40d927)\"" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" podUID="c703820f-a08c-4666-abc7-48954e40d927" Oct 09 11:08:47 crc kubenswrapper[4923]: I1009 11:08:47.601747 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:08:47 crc kubenswrapper[4923]: E1009 11:08:47.602370 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:08:47 crc kubenswrapper[4923]: I1009 11:08:47.962536 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-kclss_46e74a19-0dfc-4cac-a8ba-98062a9ca3ed/cert-manager-controller/0.log" Oct 09 11:08:48 crc kubenswrapper[4923]: I1009 11:08:48.680659 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xhmhr_5155ab6c-6e6c-4937-85cd-55fc49b7ce1a/cert-manager-webhook/0.log" Oct 09 11:08:48 crc kubenswrapper[4923]: I1009 11:08:48.730737 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-tfrn9_9aac563b-4a10-4a9e-a5c0-b3dfb5e4476d/cert-manager-cainjector/0.log" Oct 09 11:08:49 crc kubenswrapper[4923]: I1009 11:08:49.343658 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:08:49 crc kubenswrapper[4923]: I1009 11:08:49.343739 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:08:49 crc kubenswrapper[4923]: I1009 11:08:49.344803 4923 scope.go:117] "RemoveContainer" containerID="670db4a03e2a459f6e766d46d957174f0b06617bc7aeac6d8ff72e27c423a1f6" Oct 09 11:08:49 crc kubenswrapper[4923]: E1009 11:08:49.345286 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=test-operator-controller-manager-776875b78c-5975g_openstack-operators(c703820f-a08c-4666-abc7-48954e40d927)\"" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" podUID="c703820f-a08c-4666-abc7-48954e40d927" Oct 09 11:09:00 crc kubenswrapper[4923]: I1009 11:09:00.601483 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:09:00 crc kubenswrapper[4923]: E1009 11:09:00.602804 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:09:01 crc kubenswrapper[4923]: I1009 11:09:01.601240 4923 scope.go:117] "RemoveContainer" containerID="670db4a03e2a459f6e766d46d957174f0b06617bc7aeac6d8ff72e27c423a1f6" Oct 09 11:09:02 crc kubenswrapper[4923]: I1009 11:09:02.375877 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-5swk6_685e4264-fd07-4f0c-8223-7f4ac22e89b8/nmstate-console-plugin/0.log" Oct 09 11:09:02 crc kubenswrapper[4923]: I1009 11:09:02.533946 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-9t6jc_59a2d879-dfd9-4342-aef1-d99ebcbc847e/nmstate-handler/0.log" Oct 09 11:09:02 crc kubenswrapper[4923]: I1009 11:09:02.597258 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-xsz5l_9170945d-01c4-4c87-a7fa-c5220a39acdd/kube-rbac-proxy/0.log" Oct 09 11:09:02 crc kubenswrapper[4923]: I1009 11:09:02.634349 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-xsz5l_9170945d-01c4-4c87-a7fa-c5220a39acdd/nmstate-metrics/0.log" Oct 09 11:09:02 crc kubenswrapper[4923]: I1009 11:09:02.807241 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-6sg6l_fd8d3d9c-df27-445f-8ba6-5c39074945bf/nmstate-operator/0.log" Oct 09 11:09:02 crc kubenswrapper[4923]: I1009 11:09:02.859732 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerStarted","Data":"3f28694185ef989f6fdabef46dda242d2ffc517f5b7aa8a72bf8726af19f6c7f"} Oct 09 11:09:02 crc kubenswrapper[4923]: I1009 11:09:02.860088 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:09:02 crc kubenswrapper[4923]: I1009 11:09:02.878627 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-r8kkd_8e7041e7-f857-461a-a327-3a267fdfe83b/nmstate-webhook/0.log" Oct 09 11:09:09 crc kubenswrapper[4923]: I1009 11:09:09.346358 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:09:15 crc kubenswrapper[4923]: I1009 11:09:15.601416 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:09:15 crc kubenswrapper[4923]: E1009 11:09:15.602399 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.067778 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-fgmg8_70363f45-667f-4680-bf83-5358c618b73d/kube-rbac-proxy/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.230618 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-fgmg8_70363f45-667f-4680-bf83-5358c618b73d/controller/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.374016 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-frr-files/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.517263 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-metrics/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.523550 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-frr-files/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.541217 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-reloader/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.595886 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-reloader/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.842924 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-metrics/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.850211 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-reloader/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.853133 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-metrics/0.log" Oct 09 11:09:18 crc kubenswrapper[4923]: I1009 11:09:18.894293 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-frr-files/0.log" Oct 09 11:09:19 crc kubenswrapper[4923]: I1009 11:09:19.070609 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-reloader/0.log" Oct 09 11:09:19 crc kubenswrapper[4923]: I1009 11:09:19.115540 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-frr-files/0.log" Oct 09 11:09:19 crc kubenswrapper[4923]: I1009 11:09:19.115942 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/cp-metrics/0.log" Oct 09 11:09:19 crc kubenswrapper[4923]: I1009 11:09:19.128938 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/controller/0.log" Oct 09 11:09:19 crc kubenswrapper[4923]: I1009 11:09:19.408969 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/frr-metrics/0.log" Oct 09 11:09:19 crc kubenswrapper[4923]: I1009 11:09:19.418341 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/kube-rbac-proxy/0.log" Oct 09 11:09:19 crc kubenswrapper[4923]: I1009 11:09:19.443700 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/kube-rbac-proxy-frr/0.log" Oct 09 11:09:19 crc kubenswrapper[4923]: I1009 11:09:19.730888 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/reloader/0.log" Oct 09 11:09:19 crc kubenswrapper[4923]: I1009 11:09:19.738013 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-cp5k6_b15e8e44-4023-48fe-8428-7b89368ada72/frr-k8s-webhook-server/0.log" Oct 09 11:09:20 crc kubenswrapper[4923]: I1009 11:09:20.080224 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-59d99554d9-d78r5_2da4daff-10af-46a7-9b37-dde79cb66313/manager/0.log" Oct 09 11:09:20 crc kubenswrapper[4923]: I1009 11:09:20.295513 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-b6c67695c-wqcnz_711d6835-1092-41a2-9de3-476e55baf25c/webhook-server/0.log" Oct 09 11:09:20 crc kubenswrapper[4923]: I1009 11:09:20.430729 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-58t9k_e3ed7aa1-9fa9-4dd6-9701-3845073d3422/kube-rbac-proxy/0.log" Oct 09 11:09:21 crc kubenswrapper[4923]: I1009 11:09:21.120239 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-58t9k_e3ed7aa1-9fa9-4dd6-9701-3845073d3422/speaker/0.log" Oct 09 11:09:21 crc kubenswrapper[4923]: I1009 11:09:21.299294 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-vsqk8_84d4e047-944f-41c7-859c-658a76ecc3be/frr/0.log" Oct 09 11:09:28 crc kubenswrapper[4923]: I1009 11:09:28.601872 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:09:28 crc kubenswrapper[4923]: E1009 11:09:28.603313 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:09:34 crc kubenswrapper[4923]: I1009 11:09:34.997614 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn_7c8cda6e-95c6-4907-8aa0-6bbf50a0c439/util/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.186011 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn_7c8cda6e-95c6-4907-8aa0-6bbf50a0c439/util/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.204058 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn_7c8cda6e-95c6-4907-8aa0-6bbf50a0c439/pull/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.269941 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn_7c8cda6e-95c6-4907-8aa0-6bbf50a0c439/pull/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.419902 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn_7c8cda6e-95c6-4907-8aa0-6bbf50a0c439/pull/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.428606 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn_7c8cda6e-95c6-4907-8aa0-6bbf50a0c439/util/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.458329 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2vshsn_7c8cda6e-95c6-4907-8aa0-6bbf50a0c439/extract/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.617554 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcl9q_16877656-531f-42f1-9c4e-c72e99f6a536/extract-utilities/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.854059 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcl9q_16877656-531f-42f1-9c4e-c72e99f6a536/extract-utilities/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.854343 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcl9q_16877656-531f-42f1-9c4e-c72e99f6a536/extract-content/0.log" Oct 09 11:09:35 crc kubenswrapper[4923]: I1009 11:09:35.871532 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcl9q_16877656-531f-42f1-9c4e-c72e99f6a536/extract-content/0.log" Oct 09 11:09:36 crc kubenswrapper[4923]: I1009 11:09:36.099083 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcl9q_16877656-531f-42f1-9c4e-c72e99f6a536/extract-content/0.log" Oct 09 11:09:36 crc kubenswrapper[4923]: I1009 11:09:36.117041 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcl9q_16877656-531f-42f1-9c4e-c72e99f6a536/extract-utilities/0.log" Oct 09 11:09:36 crc kubenswrapper[4923]: I1009 11:09:36.366824 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brvm4_8d1e82e8-6bce-4de6-a52d-348db9bf7e6e/extract-utilities/0.log" Oct 09 11:09:36 crc kubenswrapper[4923]: I1009 11:09:36.607509 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-jcl9q_16877656-531f-42f1-9c4e-c72e99f6a536/registry-server/0.log" Oct 09 11:09:36 crc kubenswrapper[4923]: I1009 11:09:36.672795 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brvm4_8d1e82e8-6bce-4de6-a52d-348db9bf7e6e/extract-utilities/0.log" Oct 09 11:09:36 crc kubenswrapper[4923]: I1009 11:09:36.694936 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brvm4_8d1e82e8-6bce-4de6-a52d-348db9bf7e6e/extract-content/0.log" Oct 09 11:09:36 crc kubenswrapper[4923]: I1009 11:09:36.734608 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brvm4_8d1e82e8-6bce-4de6-a52d-348db9bf7e6e/extract-content/0.log" Oct 09 11:09:36 crc kubenswrapper[4923]: I1009 11:09:36.930781 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brvm4_8d1e82e8-6bce-4de6-a52d-348db9bf7e6e/extract-utilities/0.log" Oct 09 11:09:36 crc kubenswrapper[4923]: I1009 11:09:36.931421 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brvm4_8d1e82e8-6bce-4de6-a52d-348db9bf7e6e/extract-content/0.log" Oct 09 11:09:37 crc kubenswrapper[4923]: I1009 11:09:37.239651 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg_4c7a6073-930b-477d-a8af-67196a768b71/util/0.log" Oct 09 11:09:37 crc kubenswrapper[4923]: I1009 11:09:37.521732 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg_4c7a6073-930b-477d-a8af-67196a768b71/pull/0.log" Oct 09 11:09:37 crc kubenswrapper[4923]: I1009 11:09:37.524622 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg_4c7a6073-930b-477d-a8af-67196a768b71/pull/0.log" Oct 09 11:09:37 crc kubenswrapper[4923]: I1009 11:09:37.600385 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg_4c7a6073-930b-477d-a8af-67196a768b71/util/0.log" Oct 09 11:09:37 crc kubenswrapper[4923]: I1009 11:09:37.707848 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-brvm4_8d1e82e8-6bce-4de6-a52d-348db9bf7e6e/registry-server/0.log" Oct 09 11:09:37 crc kubenswrapper[4923]: I1009 11:09:37.812372 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg_4c7a6073-930b-477d-a8af-67196a768b71/util/0.log" Oct 09 11:09:37 crc kubenswrapper[4923]: I1009 11:09:37.860414 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg_4c7a6073-930b-477d-a8af-67196a768b71/extract/0.log" Oct 09 11:09:37 crc kubenswrapper[4923]: I1009 11:09:37.864662 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835cvwvxg_4c7a6073-930b-477d-a8af-67196a768b71/pull/0.log" Oct 09 11:09:38 crc kubenswrapper[4923]: I1009 11:09:38.048406 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-k9clm_a5511d1b-1e03-4020-823d-1259256d7123/marketplace-operator/0.log" Oct 09 11:09:38 crc kubenswrapper[4923]: I1009 11:09:38.083541 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-44dmh_d1d773cf-afe5-4611-a180-c027b6e3c129/extract-utilities/0.log" Oct 09 11:09:38 crc kubenswrapper[4923]: I1009 11:09:38.319067 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-44dmh_d1d773cf-afe5-4611-a180-c027b6e3c129/extract-utilities/0.log" Oct 09 11:09:38 crc kubenswrapper[4923]: I1009 11:09:38.329708 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-44dmh_d1d773cf-afe5-4611-a180-c027b6e3c129/extract-content/0.log" Oct 09 11:09:38 crc kubenswrapper[4923]: I1009 11:09:38.371292 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-44dmh_d1d773cf-afe5-4611-a180-c027b6e3c129/extract-content/0.log" Oct 09 11:09:38 crc kubenswrapper[4923]: I1009 11:09:38.549907 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-44dmh_d1d773cf-afe5-4611-a180-c027b6e3c129/extract-content/0.log" Oct 09 11:09:38 crc kubenswrapper[4923]: I1009 11:09:38.585908 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-44dmh_d1d773cf-afe5-4611-a180-c027b6e3c129/extract-utilities/0.log" Oct 09 11:09:38 crc kubenswrapper[4923]: I1009 11:09:38.803375 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-44dmh_d1d773cf-afe5-4611-a180-c027b6e3c129/registry-server/0.log" Oct 09 11:09:38 crc kubenswrapper[4923]: I1009 11:09:38.817909 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4btwb_3ddd9f92-4734-4e5a-b5cb-7d4e7448596f/extract-utilities/0.log" Oct 09 11:09:39 crc kubenswrapper[4923]: I1009 11:09:39.117806 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4btwb_3ddd9f92-4734-4e5a-b5cb-7d4e7448596f/extract-utilities/0.log" Oct 09 11:09:39 crc kubenswrapper[4923]: I1009 11:09:39.179622 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4btwb_3ddd9f92-4734-4e5a-b5cb-7d4e7448596f/extract-content/0.log" Oct 09 11:09:39 crc kubenswrapper[4923]: I1009 11:09:39.294881 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4btwb_3ddd9f92-4734-4e5a-b5cb-7d4e7448596f/extract-content/0.log" Oct 09 11:09:39 crc kubenswrapper[4923]: I1009 11:09:39.367564 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4btwb_3ddd9f92-4734-4e5a-b5cb-7d4e7448596f/extract-utilities/0.log" Oct 09 11:09:39 crc kubenswrapper[4923]: I1009 11:09:39.425605 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4btwb_3ddd9f92-4734-4e5a-b5cb-7d4e7448596f/extract-content/0.log" Oct 09 11:09:39 crc kubenswrapper[4923]: I1009 11:09:39.945207 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-4btwb_3ddd9f92-4734-4e5a-b5cb-7d4e7448596f/registry-server/0.log" Oct 09 11:09:43 crc kubenswrapper[4923]: I1009 11:09:43.603329 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:09:43 crc kubenswrapper[4923]: E1009 11:09:43.604500 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:09:54 crc kubenswrapper[4923]: I1009 11:09:54.608403 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:09:54 crc kubenswrapper[4923]: E1009 11:09:54.609435 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-frh4j_openshift-machine-config-operator(1130e920-d4a8-44fe-8bb4-050213d1cff2)\"" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.415528 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9wqdz"] Oct 09 11:09:57 crc kubenswrapper[4923]: E1009 11:09:57.416510 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerName="extract-utilities" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.416527 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerName="extract-utilities" Oct 09 11:09:57 crc kubenswrapper[4923]: E1009 11:09:57.416543 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerName="registry-server" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.416549 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerName="registry-server" Oct 09 11:09:57 crc kubenswrapper[4923]: E1009 11:09:57.416561 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerName="extract-content" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.416568 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerName="extract-content" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.416806 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="b74cbadd-ac6e-49a5-ab5d-78fd3f344cdc" containerName="registry-server" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.418278 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.441391 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wqdz"] Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.499969 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-catalog-content\") pod \"community-operators-9wqdz\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.500474 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl689\" (UniqueName: \"kubernetes.io/projected/27453a3d-6076-4505-91bd-f93e879cd7db-kube-api-access-xl689\") pod \"community-operators-9wqdz\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.500523 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-utilities\") pod \"community-operators-9wqdz\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.602218 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-catalog-content\") pod \"community-operators-9wqdz\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.602347 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl689\" (UniqueName: \"kubernetes.io/projected/27453a3d-6076-4505-91bd-f93e879cd7db-kube-api-access-xl689\") pod \"community-operators-9wqdz\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.602389 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-utilities\") pod \"community-operators-9wqdz\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.602781 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-catalog-content\") pod \"community-operators-9wqdz\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:57 crc kubenswrapper[4923]: I1009 11:09:57.602978 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-utilities\") pod \"community-operators-9wqdz\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:58 crc kubenswrapper[4923]: I1009 11:09:58.170049 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl689\" (UniqueName: \"kubernetes.io/projected/27453a3d-6076-4505-91bd-f93e879cd7db-kube-api-access-xl689\") pod \"community-operators-9wqdz\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:58 crc kubenswrapper[4923]: I1009 11:09:58.386673 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:09:58 crc kubenswrapper[4923]: I1009 11:09:58.976896 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9wqdz"] Oct 09 11:09:59 crc kubenswrapper[4923]: I1009 11:09:59.473529 4923 generic.go:334] "Generic (PLEG): container finished" podID="27453a3d-6076-4505-91bd-f93e879cd7db" containerID="026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077" exitCode=0 Oct 09 11:09:59 crc kubenswrapper[4923]: I1009 11:09:59.474032 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wqdz" event={"ID":"27453a3d-6076-4505-91bd-f93e879cd7db","Type":"ContainerDied","Data":"026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077"} Oct 09 11:09:59 crc kubenswrapper[4923]: I1009 11:09:59.474073 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wqdz" event={"ID":"27453a3d-6076-4505-91bd-f93e879cd7db","Type":"ContainerStarted","Data":"c98adfcf27b629d5e589475f502c9c5bbc7baa91314499d0a8162e7c6e996c44"} Oct 09 11:09:59 crc kubenswrapper[4923]: I1009 11:09:59.476711 4923 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 11:10:01 crc kubenswrapper[4923]: I1009 11:10:01.516746 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wqdz" event={"ID":"27453a3d-6076-4505-91bd-f93e879cd7db","Type":"ContainerStarted","Data":"5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143"} Oct 09 11:10:02 crc kubenswrapper[4923]: I1009 11:10:02.532797 4923 generic.go:334] "Generic (PLEG): container finished" podID="27453a3d-6076-4505-91bd-f93e879cd7db" containerID="5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143" exitCode=0 Oct 09 11:10:02 crc kubenswrapper[4923]: I1009 11:10:02.532942 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wqdz" event={"ID":"27453a3d-6076-4505-91bd-f93e879cd7db","Type":"ContainerDied","Data":"5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143"} Oct 09 11:10:03 crc kubenswrapper[4923]: I1009 11:10:03.548427 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wqdz" event={"ID":"27453a3d-6076-4505-91bd-f93e879cd7db","Type":"ContainerStarted","Data":"d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f"} Oct 09 11:10:03 crc kubenswrapper[4923]: I1009 11:10:03.572152 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9wqdz" podStartSLOduration=2.962727892 podStartE2EDuration="6.572126458s" podCreationTimestamp="2025-10-09 11:09:57 +0000 UTC" firstStartedPulling="2025-10-09 11:09:59.476437089 +0000 UTC m=+3885.544618845" lastFinishedPulling="2025-10-09 11:10:03.085835655 +0000 UTC m=+3889.154017411" observedRunningTime="2025-10-09 11:10:03.569252378 +0000 UTC m=+3889.637434134" watchObservedRunningTime="2025-10-09 11:10:03.572126458 +0000 UTC m=+3889.640308214" Oct 09 11:10:08 crc kubenswrapper[4923]: I1009 11:10:08.386929 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:10:08 crc kubenswrapper[4923]: I1009 11:10:08.389423 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:10:08 crc kubenswrapper[4923]: I1009 11:10:08.601943 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:10:08 crc kubenswrapper[4923]: I1009 11:10:08.613808 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:10:09 crc kubenswrapper[4923]: I1009 11:10:09.609210 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"74f7108ed6c01c53d833bedaa32717bf22d3e36d45219cfd30941fc6c77d8a25"} Oct 09 11:10:09 crc kubenswrapper[4923]: I1009 11:10:09.707005 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:10:09 crc kubenswrapper[4923]: I1009 11:10:09.770544 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wqdz"] Oct 09 11:10:11 crc kubenswrapper[4923]: I1009 11:10:11.651365 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9wqdz" podUID="27453a3d-6076-4505-91bd-f93e879cd7db" containerName="registry-server" containerID="cri-o://d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f" gracePeriod=2 Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.334984 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.534325 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl689\" (UniqueName: \"kubernetes.io/projected/27453a3d-6076-4505-91bd-f93e879cd7db-kube-api-access-xl689\") pod \"27453a3d-6076-4505-91bd-f93e879cd7db\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.534433 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-utilities\") pod \"27453a3d-6076-4505-91bd-f93e879cd7db\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.534482 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-catalog-content\") pod \"27453a3d-6076-4505-91bd-f93e879cd7db\" (UID: \"27453a3d-6076-4505-91bd-f93e879cd7db\") " Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.536732 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-utilities" (OuterVolumeSpecName: "utilities") pod "27453a3d-6076-4505-91bd-f93e879cd7db" (UID: "27453a3d-6076-4505-91bd-f93e879cd7db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.554745 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27453a3d-6076-4505-91bd-f93e879cd7db-kube-api-access-xl689" (OuterVolumeSpecName: "kube-api-access-xl689") pod "27453a3d-6076-4505-91bd-f93e879cd7db" (UID: "27453a3d-6076-4505-91bd-f93e879cd7db"). InnerVolumeSpecName "kube-api-access-xl689". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.635613 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27453a3d-6076-4505-91bd-f93e879cd7db" (UID: "27453a3d-6076-4505-91bd-f93e879cd7db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.639379 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl689\" (UniqueName: \"kubernetes.io/projected/27453a3d-6076-4505-91bd-f93e879cd7db-kube-api-access-xl689\") on node \"crc\" DevicePath \"\"" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.639422 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.639437 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27453a3d-6076-4505-91bd-f93e879cd7db-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.696035 4923 generic.go:334] "Generic (PLEG): container finished" podID="27453a3d-6076-4505-91bd-f93e879cd7db" containerID="d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f" exitCode=0 Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.696101 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wqdz" event={"ID":"27453a3d-6076-4505-91bd-f93e879cd7db","Type":"ContainerDied","Data":"d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f"} Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.696147 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9wqdz" event={"ID":"27453a3d-6076-4505-91bd-f93e879cd7db","Type":"ContainerDied","Data":"c98adfcf27b629d5e589475f502c9c5bbc7baa91314499d0a8162e7c6e996c44"} Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.696172 4923 scope.go:117] "RemoveContainer" containerID="d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.696276 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9wqdz" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.768019 4923 scope.go:117] "RemoveContainer" containerID="5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.769540 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9wqdz"] Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.784046 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9wqdz"] Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.837031 4923 scope.go:117] "RemoveContainer" containerID="026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.892556 4923 scope.go:117] "RemoveContainer" containerID="d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f" Oct 09 11:10:12 crc kubenswrapper[4923]: E1009 11:10:12.895478 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f\": container with ID starting with d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f not found: ID does not exist" containerID="d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.895547 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f"} err="failed to get container status \"d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f\": rpc error: code = NotFound desc = could not find container \"d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f\": container with ID starting with d4740e58e381180be39559796d644ed5e2bb23ea853fbd815b07b80c6223161f not found: ID does not exist" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.895592 4923 scope.go:117] "RemoveContainer" containerID="5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143" Oct 09 11:10:12 crc kubenswrapper[4923]: E1009 11:10:12.898137 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143\": container with ID starting with 5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143 not found: ID does not exist" containerID="5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.898190 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143"} err="failed to get container status \"5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143\": rpc error: code = NotFound desc = could not find container \"5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143\": container with ID starting with 5410ebf023087c083ee47de39afbd693117855b0c595de210f43f4246183d143 not found: ID does not exist" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.898224 4923 scope.go:117] "RemoveContainer" containerID="026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077" Oct 09 11:10:12 crc kubenswrapper[4923]: E1009 11:10:12.898937 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077\": container with ID starting with 026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077 not found: ID does not exist" containerID="026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077" Oct 09 11:10:12 crc kubenswrapper[4923]: I1009 11:10:12.898998 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077"} err="failed to get container status \"026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077\": rpc error: code = NotFound desc = could not find container \"026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077\": container with ID starting with 026c3578cb39ffd33bbc18ac626d554032e43d10cfb43e6c1c8b8154aace6077 not found: ID does not exist" Oct 09 11:10:14 crc kubenswrapper[4923]: I1009 11:10:14.630967 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27453a3d-6076-4505-91bd-f93e879cd7db" path="/var/lib/kubelet/pods/27453a3d-6076-4505-91bd-f93e879cd7db/volumes" Oct 09 11:10:21 crc kubenswrapper[4923]: I1009 11:10:21.075742 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-mxv9k"] Oct 09 11:10:21 crc kubenswrapper[4923]: I1009 11:10:21.088386 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-mxv9k"] Oct 09 11:10:22 crc kubenswrapper[4923]: I1009 11:10:22.622017 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f093ece7-e8f5-4beb-8f89-612afece0428" path="/var/lib/kubelet/pods/f093ece7-e8f5-4beb-8f89-612afece0428/volumes" Oct 09 11:10:32 crc kubenswrapper[4923]: I1009 11:10:32.032867 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-37f2-account-create-64wqp"] Oct 09 11:10:32 crc kubenswrapper[4923]: I1009 11:10:32.043597 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-37f2-account-create-64wqp"] Oct 09 11:10:32 crc kubenswrapper[4923]: I1009 11:10:32.619885 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53939d29-b8dc-4a59-a7b7-37a8675ced9a" path="/var/lib/kubelet/pods/53939d29-b8dc-4a59-a7b7-37a8675ced9a/volumes" Oct 09 11:10:52 crc kubenswrapper[4923]: I1009 11:10:52.062189 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-z6tpv"] Oct 09 11:10:52 crc kubenswrapper[4923]: I1009 11:10:52.071695 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-z6tpv"] Oct 09 11:10:52 crc kubenswrapper[4923]: I1009 11:10:52.616218 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e665c786-66db-460a-9bd8-c1ec9ecfb978" path="/var/lib/kubelet/pods/e665c786-66db-460a-9bd8-c1ec9ecfb978/volumes" Oct 09 11:11:21 crc kubenswrapper[4923]: I1009 11:11:21.020627 4923 scope.go:117] "RemoveContainer" containerID="3631d6f320264077077990d367e92e333d2e99d4728f0911dfa11666d3ca8a2a" Oct 09 11:11:21 crc kubenswrapper[4923]: I1009 11:11:21.052470 4923 scope.go:117] "RemoveContainer" containerID="201576e47d13293de82c7debb17e5b304dc5a56b525ae1208a07029a1b550437" Oct 09 11:11:21 crc kubenswrapper[4923]: I1009 11:11:21.100916 4923 scope.go:117] "RemoveContainer" containerID="5ae02b6345a571f707ec2685df5c5510d77e8f74bb5367b11aa75514e34aaf8d" Oct 09 11:11:36 crc kubenswrapper[4923]: I1009 11:11:36.586827 4923 generic.go:334] "Generic (PLEG): container finished" podID="c703820f-a08c-4666-abc7-48954e40d927" containerID="3f28694185ef989f6fdabef46dda242d2ffc517f5b7aa8a72bf8726af19f6c7f" exitCode=1 Oct 09 11:11:36 crc kubenswrapper[4923]: I1009 11:11:36.587555 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerDied","Data":"3f28694185ef989f6fdabef46dda242d2ffc517f5b7aa8a72bf8726af19f6c7f"} Oct 09 11:11:36 crc kubenswrapper[4923]: I1009 11:11:36.587603 4923 scope.go:117] "RemoveContainer" containerID="670db4a03e2a459f6e766d46d957174f0b06617bc7aeac6d8ff72e27c423a1f6" Oct 09 11:11:36 crc kubenswrapper[4923]: I1009 11:11:36.588644 4923 scope.go:117] "RemoveContainer" containerID="3f28694185ef989f6fdabef46dda242d2ffc517f5b7aa8a72bf8726af19f6c7f" Oct 09 11:11:36 crc kubenswrapper[4923]: E1009 11:11:36.589018 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=test-operator-controller-manager-776875b78c-5975g_openstack-operators(c703820f-a08c-4666-abc7-48954e40d927)\"" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" podUID="c703820f-a08c-4666-abc7-48954e40d927" Oct 09 11:11:39 crc kubenswrapper[4923]: I1009 11:11:39.343882 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:11:39 crc kubenswrapper[4923]: I1009 11:11:39.347047 4923 scope.go:117] "RemoveContainer" containerID="3f28694185ef989f6fdabef46dda242d2ffc517f5b7aa8a72bf8726af19f6c7f" Oct 09 11:11:39 crc kubenswrapper[4923]: E1009 11:11:39.347441 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=test-operator-controller-manager-776875b78c-5975g_openstack-operators(c703820f-a08c-4666-abc7-48954e40d927)\"" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" podUID="c703820f-a08c-4666-abc7-48954e40d927" Oct 09 11:11:40 crc kubenswrapper[4923]: I1009 11:11:40.638663 4923 generic.go:334] "Generic (PLEG): container finished" podID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" containerID="365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e" exitCode=0 Oct 09 11:11:40 crc kubenswrapper[4923]: I1009 11:11:40.638797 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" event={"ID":"2db7ad3b-f634-4a87-a5af-f4d00b913e85","Type":"ContainerDied","Data":"365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e"} Oct 09 11:11:40 crc kubenswrapper[4923]: I1009 11:11:40.640289 4923 scope.go:117] "RemoveContainer" containerID="365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e" Oct 09 11:11:41 crc kubenswrapper[4923]: I1009 11:11:41.438613 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-p2kmf_must-gather-x5cfb_2db7ad3b-f634-4a87-a5af-f4d00b913e85/gather/0.log" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.027626 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-p2kmf/must-gather-x5cfb"] Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.028771 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" podUID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" containerName="copy" containerID="cri-o://a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319" gracePeriod=2 Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.038189 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-p2kmf/must-gather-x5cfb"] Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.343553 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.350015 4923 scope.go:117] "RemoveContainer" containerID="3f28694185ef989f6fdabef46dda242d2ffc517f5b7aa8a72bf8726af19f6c7f" Oct 09 11:11:49 crc kubenswrapper[4923]: E1009 11:11:49.350625 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=test-operator-controller-manager-776875b78c-5975g_openstack-operators(c703820f-a08c-4666-abc7-48954e40d927)\"" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" podUID="c703820f-a08c-4666-abc7-48954e40d927" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.576450 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-p2kmf_must-gather-x5cfb_2db7ad3b-f634-4a87-a5af-f4d00b913e85/copy/0.log" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.577185 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.728962 4923 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-p2kmf_must-gather-x5cfb_2db7ad3b-f634-4a87-a5af-f4d00b913e85/copy/0.log" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.729582 4923 generic.go:334] "Generic (PLEG): container finished" podID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" containerID="a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319" exitCode=143 Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.729660 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-p2kmf/must-gather-x5cfb" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.729670 4923 scope.go:117] "RemoveContainer" containerID="a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.755867 4923 scope.go:117] "RemoveContainer" containerID="365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.781431 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shfgd\" (UniqueName: \"kubernetes.io/projected/2db7ad3b-f634-4a87-a5af-f4d00b913e85-kube-api-access-shfgd\") pod \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\" (UID: \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\") " Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.781561 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2db7ad3b-f634-4a87-a5af-f4d00b913e85-must-gather-output\") pod \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\" (UID: \"2db7ad3b-f634-4a87-a5af-f4d00b913e85\") " Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.793201 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2db7ad3b-f634-4a87-a5af-f4d00b913e85-kube-api-access-shfgd" (OuterVolumeSpecName: "kube-api-access-shfgd") pod "2db7ad3b-f634-4a87-a5af-f4d00b913e85" (UID: "2db7ad3b-f634-4a87-a5af-f4d00b913e85"). InnerVolumeSpecName "kube-api-access-shfgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.863337 4923 scope.go:117] "RemoveContainer" containerID="a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319" Oct 09 11:11:49 crc kubenswrapper[4923]: E1009 11:11:49.864001 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319\": container with ID starting with a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319 not found: ID does not exist" containerID="a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.864045 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319"} err="failed to get container status \"a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319\": rpc error: code = NotFound desc = could not find container \"a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319\": container with ID starting with a95cb76499f6b6d376472ca5a26ea0e8ddc00034d20b62bf3b71cdd894368319 not found: ID does not exist" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.864079 4923 scope.go:117] "RemoveContainer" containerID="365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e" Oct 09 11:11:49 crc kubenswrapper[4923]: E1009 11:11:49.864709 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e\": container with ID starting with 365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e not found: ID does not exist" containerID="365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.864789 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e"} err="failed to get container status \"365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e\": rpc error: code = NotFound desc = could not find container \"365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e\": container with ID starting with 365ea7c7d5d0a4bacd07c9b8ca654b33eada6ed4a19acd73a28c9b1f3bf2e24e not found: ID does not exist" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.887707 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shfgd\" (UniqueName: \"kubernetes.io/projected/2db7ad3b-f634-4a87-a5af-f4d00b913e85-kube-api-access-shfgd\") on node \"crc\" DevicePath \"\"" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.946545 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2db7ad3b-f634-4a87-a5af-f4d00b913e85-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2db7ad3b-f634-4a87-a5af-f4d00b913e85" (UID: "2db7ad3b-f634-4a87-a5af-f4d00b913e85"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:11:49 crc kubenswrapper[4923]: I1009 11:11:49.990372 4923 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2db7ad3b-f634-4a87-a5af-f4d00b913e85-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 09 11:11:50 crc kubenswrapper[4923]: I1009 11:11:50.613103 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" path="/var/lib/kubelet/pods/2db7ad3b-f634-4a87-a5af-f4d00b913e85/volumes" Oct 09 11:12:03 crc kubenswrapper[4923]: I1009 11:12:03.602269 4923 scope.go:117] "RemoveContainer" containerID="3f28694185ef989f6fdabef46dda242d2ffc517f5b7aa8a72bf8726af19f6c7f" Oct 09 11:12:03 crc kubenswrapper[4923]: I1009 11:12:03.877379 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerStarted","Data":"417e35d6ac99509c97045327b4c40105b78ebccbf3ac04a6ea482e8989638a07"} Oct 09 11:12:03 crc kubenswrapper[4923]: I1009 11:12:03.879346 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:12:09 crc kubenswrapper[4923]: I1009 11:12:09.346731 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:12:24 crc kubenswrapper[4923]: I1009 11:12:24.606386 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 11:12:24 crc kubenswrapper[4923]: I1009 11:12:24.607235 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 11:12:54 crc kubenswrapper[4923]: I1009 11:12:54.600026 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 11:12:54 crc kubenswrapper[4923]: I1009 11:12:54.600909 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.775504 4923 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-s67nc"] Oct 09 11:13:00 crc kubenswrapper[4923]: E1009 11:13:00.776816 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27453a3d-6076-4505-91bd-f93e879cd7db" containerName="registry-server" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.776834 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="27453a3d-6076-4505-91bd-f93e879cd7db" containerName="registry-server" Oct 09 11:13:00 crc kubenswrapper[4923]: E1009 11:13:00.776876 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" containerName="copy" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.776885 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" containerName="copy" Oct 09 11:13:00 crc kubenswrapper[4923]: E1009 11:13:00.776902 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27453a3d-6076-4505-91bd-f93e879cd7db" containerName="extract-content" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.776909 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="27453a3d-6076-4505-91bd-f93e879cd7db" containerName="extract-content" Oct 09 11:13:00 crc kubenswrapper[4923]: E1009 11:13:00.776927 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" containerName="gather" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.776934 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" containerName="gather" Oct 09 11:13:00 crc kubenswrapper[4923]: E1009 11:13:00.776945 4923 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27453a3d-6076-4505-91bd-f93e879cd7db" containerName="extract-utilities" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.776951 4923 state_mem.go:107] "Deleted CPUSet assignment" podUID="27453a3d-6076-4505-91bd-f93e879cd7db" containerName="extract-utilities" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.777199 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="27453a3d-6076-4505-91bd-f93e879cd7db" containerName="registry-server" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.777258 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" containerName="copy" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.777289 4923 memory_manager.go:354] "RemoveStaleState removing state" podUID="2db7ad3b-f634-4a87-a5af-f4d00b913e85" containerName="gather" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.781778 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.792688 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s67nc"] Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.958555 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-utilities\") pod \"redhat-marketplace-s67nc\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.958992 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw8nl\" (UniqueName: \"kubernetes.io/projected/64864284-412f-4b25-b8ca-e8284b7b0173-kube-api-access-hw8nl\") pod \"redhat-marketplace-s67nc\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:00 crc kubenswrapper[4923]: I1009 11:13:00.959046 4923 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-catalog-content\") pod \"redhat-marketplace-s67nc\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:01 crc kubenswrapper[4923]: I1009 11:13:01.060723 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw8nl\" (UniqueName: \"kubernetes.io/projected/64864284-412f-4b25-b8ca-e8284b7b0173-kube-api-access-hw8nl\") pod \"redhat-marketplace-s67nc\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:01 crc kubenswrapper[4923]: I1009 11:13:01.060894 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-catalog-content\") pod \"redhat-marketplace-s67nc\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:01 crc kubenswrapper[4923]: I1009 11:13:01.060945 4923 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-utilities\") pod \"redhat-marketplace-s67nc\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:01 crc kubenswrapper[4923]: I1009 11:13:01.061492 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-utilities\") pod \"redhat-marketplace-s67nc\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:01 crc kubenswrapper[4923]: I1009 11:13:01.061601 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-catalog-content\") pod \"redhat-marketplace-s67nc\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:01 crc kubenswrapper[4923]: I1009 11:13:01.085940 4923 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw8nl\" (UniqueName: \"kubernetes.io/projected/64864284-412f-4b25-b8ca-e8284b7b0173-kube-api-access-hw8nl\") pod \"redhat-marketplace-s67nc\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:01 crc kubenswrapper[4923]: I1009 11:13:01.115452 4923 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:01 crc kubenswrapper[4923]: I1009 11:13:01.620152 4923 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-s67nc"] Oct 09 11:13:02 crc kubenswrapper[4923]: I1009 11:13:02.440927 4923 generic.go:334] "Generic (PLEG): container finished" podID="64864284-412f-4b25-b8ca-e8284b7b0173" containerID="c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95" exitCode=0 Oct 09 11:13:02 crc kubenswrapper[4923]: I1009 11:13:02.441164 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s67nc" event={"ID":"64864284-412f-4b25-b8ca-e8284b7b0173","Type":"ContainerDied","Data":"c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95"} Oct 09 11:13:02 crc kubenswrapper[4923]: I1009 11:13:02.441321 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s67nc" event={"ID":"64864284-412f-4b25-b8ca-e8284b7b0173","Type":"ContainerStarted","Data":"cd67258ceb1e8e8e8574547f46425882043cfc1bb650eb66541960eb73d9342b"} Oct 09 11:13:03 crc kubenswrapper[4923]: I1009 11:13:03.453098 4923 generic.go:334] "Generic (PLEG): container finished" podID="64864284-412f-4b25-b8ca-e8284b7b0173" containerID="5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b" exitCode=0 Oct 09 11:13:03 crc kubenswrapper[4923]: I1009 11:13:03.453276 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s67nc" event={"ID":"64864284-412f-4b25-b8ca-e8284b7b0173","Type":"ContainerDied","Data":"5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b"} Oct 09 11:13:04 crc kubenswrapper[4923]: I1009 11:13:04.468011 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s67nc" event={"ID":"64864284-412f-4b25-b8ca-e8284b7b0173","Type":"ContainerStarted","Data":"4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677"} Oct 09 11:13:04 crc kubenswrapper[4923]: I1009 11:13:04.507505 4923 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-s67nc" podStartSLOduration=2.764740902 podStartE2EDuration="4.507474206s" podCreationTimestamp="2025-10-09 11:13:00 +0000 UTC" firstStartedPulling="2025-10-09 11:13:02.443629513 +0000 UTC m=+4068.511811269" lastFinishedPulling="2025-10-09 11:13:04.186362817 +0000 UTC m=+4070.254544573" observedRunningTime="2025-10-09 11:13:04.498428273 +0000 UTC m=+4070.566610029" watchObservedRunningTime="2025-10-09 11:13:04.507474206 +0000 UTC m=+4070.575655962" Oct 09 11:13:11 crc kubenswrapper[4923]: I1009 11:13:11.116284 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:11 crc kubenswrapper[4923]: I1009 11:13:11.117105 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:11 crc kubenswrapper[4923]: I1009 11:13:11.168733 4923 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:11 crc kubenswrapper[4923]: I1009 11:13:11.609662 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:11 crc kubenswrapper[4923]: I1009 11:13:11.662534 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s67nc"] Oct 09 11:13:13 crc kubenswrapper[4923]: I1009 11:13:13.573904 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-s67nc" podUID="64864284-412f-4b25-b8ca-e8284b7b0173" containerName="registry-server" containerID="cri-o://4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677" gracePeriod=2 Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.140423 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.304738 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-utilities\") pod \"64864284-412f-4b25-b8ca-e8284b7b0173\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.304998 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw8nl\" (UniqueName: \"kubernetes.io/projected/64864284-412f-4b25-b8ca-e8284b7b0173-kube-api-access-hw8nl\") pod \"64864284-412f-4b25-b8ca-e8284b7b0173\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.305035 4923 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-catalog-content\") pod \"64864284-412f-4b25-b8ca-e8284b7b0173\" (UID: \"64864284-412f-4b25-b8ca-e8284b7b0173\") " Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.307313 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-utilities" (OuterVolumeSpecName: "utilities") pod "64864284-412f-4b25-b8ca-e8284b7b0173" (UID: "64864284-412f-4b25-b8ca-e8284b7b0173"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.317308 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64864284-412f-4b25-b8ca-e8284b7b0173-kube-api-access-hw8nl" (OuterVolumeSpecName: "kube-api-access-hw8nl") pod "64864284-412f-4b25-b8ca-e8284b7b0173" (UID: "64864284-412f-4b25-b8ca-e8284b7b0173"). InnerVolumeSpecName "kube-api-access-hw8nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.325989 4923 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "64864284-412f-4b25-b8ca-e8284b7b0173" (UID: "64864284-412f-4b25-b8ca-e8284b7b0173"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.408857 4923 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw8nl\" (UniqueName: \"kubernetes.io/projected/64864284-412f-4b25-b8ca-e8284b7b0173-kube-api-access-hw8nl\") on node \"crc\" DevicePath \"\"" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.408900 4923 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.408914 4923 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/64864284-412f-4b25-b8ca-e8284b7b0173-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.594348 4923 generic.go:334] "Generic (PLEG): container finished" podID="64864284-412f-4b25-b8ca-e8284b7b0173" containerID="4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677" exitCode=0 Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.594422 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s67nc" event={"ID":"64864284-412f-4b25-b8ca-e8284b7b0173","Type":"ContainerDied","Data":"4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677"} Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.594458 4923 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-s67nc" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.594472 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-s67nc" event={"ID":"64864284-412f-4b25-b8ca-e8284b7b0173","Type":"ContainerDied","Data":"cd67258ceb1e8e8e8574547f46425882043cfc1bb650eb66541960eb73d9342b"} Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.594502 4923 scope.go:117] "RemoveContainer" containerID="4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.628127 4923 scope.go:117] "RemoveContainer" containerID="5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.657849 4923 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-s67nc"] Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.669669 4923 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-s67nc"] Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.673099 4923 scope.go:117] "RemoveContainer" containerID="c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.747522 4923 scope.go:117] "RemoveContainer" containerID="4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677" Oct 09 11:13:14 crc kubenswrapper[4923]: E1009 11:13:14.748377 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677\": container with ID starting with 4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677 not found: ID does not exist" containerID="4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.748451 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677"} err="failed to get container status \"4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677\": rpc error: code = NotFound desc = could not find container \"4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677\": container with ID starting with 4c1bc74b53e1042e6141ea3c35f375b173dd575c2bc1f38bb296e42a7d065677 not found: ID does not exist" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.748492 4923 scope.go:117] "RemoveContainer" containerID="5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b" Oct 09 11:13:14 crc kubenswrapper[4923]: E1009 11:13:14.748957 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b\": container with ID starting with 5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b not found: ID does not exist" containerID="5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.749000 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b"} err="failed to get container status \"5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b\": rpc error: code = NotFound desc = could not find container \"5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b\": container with ID starting with 5a2b7c078aa191b3b70cb607a1335d5b39b46ca092a9e35caf3c27fa6485f65b not found: ID does not exist" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.749036 4923 scope.go:117] "RemoveContainer" containerID="c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95" Oct 09 11:13:14 crc kubenswrapper[4923]: E1009 11:13:14.749407 4923 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95\": container with ID starting with c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95 not found: ID does not exist" containerID="c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95" Oct 09 11:13:14 crc kubenswrapper[4923]: I1009 11:13:14.749442 4923 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95"} err="failed to get container status \"c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95\": rpc error: code = NotFound desc = could not find container \"c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95\": container with ID starting with c5706df187f8b116d95f32cb9a83167e4833b949a76cda8af7ea14dbcf8bfb95 not found: ID does not exist" Oct 09 11:13:16 crc kubenswrapper[4923]: E1009 11:13:16.087948 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice/crio-cd67258ceb1e8e8e8574547f46425882043cfc1bb650eb66541960eb73d9342b\": RecentStats: unable to find data in memory cache]" Oct 09 11:13:16 crc kubenswrapper[4923]: I1009 11:13:16.632095 4923 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64864284-412f-4b25-b8ca-e8284b7b0173" path="/var/lib/kubelet/pods/64864284-412f-4b25-b8ca-e8284b7b0173/volumes" Oct 09 11:13:24 crc kubenswrapper[4923]: I1009 11:13:24.599529 4923 patch_prober.go:28] interesting pod/machine-config-daemon-frh4j container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 11:13:24 crc kubenswrapper[4923]: I1009 11:13:24.600307 4923 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 11:13:24 crc kubenswrapper[4923]: I1009 11:13:24.600377 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" Oct 09 11:13:24 crc kubenswrapper[4923]: I1009 11:13:24.601427 4923 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"74f7108ed6c01c53d833bedaa32717bf22d3e36d45219cfd30941fc6c77d8a25"} pod="openshift-machine-config-operator/machine-config-daemon-frh4j" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 11:13:24 crc kubenswrapper[4923]: I1009 11:13:24.601512 4923 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" podUID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerName="machine-config-daemon" containerID="cri-o://74f7108ed6c01c53d833bedaa32717bf22d3e36d45219cfd30941fc6c77d8a25" gracePeriod=600 Oct 09 11:13:25 crc kubenswrapper[4923]: I1009 11:13:25.737034 4923 generic.go:334] "Generic (PLEG): container finished" podID="1130e920-d4a8-44fe-8bb4-050213d1cff2" containerID="74f7108ed6c01c53d833bedaa32717bf22d3e36d45219cfd30941fc6c77d8a25" exitCode=0 Oct 09 11:13:25 crc kubenswrapper[4923]: I1009 11:13:25.737287 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerDied","Data":"74f7108ed6c01c53d833bedaa32717bf22d3e36d45219cfd30941fc6c77d8a25"} Oct 09 11:13:25 crc kubenswrapper[4923]: I1009 11:13:25.741360 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-frh4j" event={"ID":"1130e920-d4a8-44fe-8bb4-050213d1cff2","Type":"ContainerStarted","Data":"422907838a643c42ba30ceab19ac6eb29ac586fc7ce96824627379c064d64903"} Oct 09 11:13:25 crc kubenswrapper[4923]: I1009 11:13:25.741392 4923 scope.go:117] "RemoveContainer" containerID="8a662afd9259f25477ff825cc73459a62df60e82e50d1a714b7c614c6efcf397" Oct 09 11:13:26 crc kubenswrapper[4923]: E1009 11:13:26.364233 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice/crio-cd67258ceb1e8e8e8574547f46425882043cfc1bb650eb66541960eb73d9342b\": RecentStats: unable to find data in memory cache]" Oct 09 11:13:36 crc kubenswrapper[4923]: E1009 11:13:36.636894 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice/crio-cd67258ceb1e8e8e8574547f46425882043cfc1bb650eb66541960eb73d9342b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice\": RecentStats: unable to find data in memory cache]" Oct 09 11:13:46 crc kubenswrapper[4923]: E1009 11:13:46.945972 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice/crio-cd67258ceb1e8e8e8574547f46425882043cfc1bb650eb66541960eb73d9342b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice\": RecentStats: unable to find data in memory cache]" Oct 09 11:13:57 crc kubenswrapper[4923]: E1009 11:13:57.238192 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice/crio-cd67258ceb1e8e8e8574547f46425882043cfc1bb650eb66541960eb73d9342b\": RecentStats: unable to find data in memory cache]" Oct 09 11:14:07 crc kubenswrapper[4923]: E1009 11:14:07.541341 4923 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice/crio-cd67258ceb1e8e8e8574547f46425882043cfc1bb650eb66541960eb73d9342b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64864284_412f_4b25_b8ca_e8284b7b0173.slice\": RecentStats: unable to find data in memory cache]" Oct 09 11:14:36 crc kubenswrapper[4923]: I1009 11:14:36.487023 4923 generic.go:334] "Generic (PLEG): container finished" podID="c703820f-a08c-4666-abc7-48954e40d927" containerID="417e35d6ac99509c97045327b4c40105b78ebccbf3ac04a6ea482e8989638a07" exitCode=1 Oct 09 11:14:36 crc kubenswrapper[4923]: I1009 11:14:36.487111 4923 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" event={"ID":"c703820f-a08c-4666-abc7-48954e40d927","Type":"ContainerDied","Data":"417e35d6ac99509c97045327b4c40105b78ebccbf3ac04a6ea482e8989638a07"} Oct 09 11:14:36 crc kubenswrapper[4923]: I1009 11:14:36.488003 4923 scope.go:117] "RemoveContainer" containerID="3f28694185ef989f6fdabef46dda242d2ffc517f5b7aa8a72bf8726af19f6c7f" Oct 09 11:14:36 crc kubenswrapper[4923]: I1009 11:14:36.489416 4923 scope.go:117] "RemoveContainer" containerID="417e35d6ac99509c97045327b4c40105b78ebccbf3ac04a6ea482e8989638a07" Oct 09 11:14:36 crc kubenswrapper[4923]: E1009 11:14:36.490805 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=test-operator-controller-manager-776875b78c-5975g_openstack-operators(c703820f-a08c-4666-abc7-48954e40d927)\"" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" podUID="c703820f-a08c-4666-abc7-48954e40d927" Oct 09 11:14:39 crc kubenswrapper[4923]: I1009 11:14:39.343850 4923 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:14:39 crc kubenswrapper[4923]: I1009 11:14:39.345318 4923 scope.go:117] "RemoveContainer" containerID="417e35d6ac99509c97045327b4c40105b78ebccbf3ac04a6ea482e8989638a07" Oct 09 11:14:39 crc kubenswrapper[4923]: E1009 11:14:39.345592 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=test-operator-controller-manager-776875b78c-5975g_openstack-operators(c703820f-a08c-4666-abc7-48954e40d927)\"" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" podUID="c703820f-a08c-4666-abc7-48954e40d927" Oct 09 11:14:49 crc kubenswrapper[4923]: I1009 11:14:49.343408 4923 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" Oct 09 11:14:49 crc kubenswrapper[4923]: I1009 11:14:49.345070 4923 scope.go:117] "RemoveContainer" containerID="417e35d6ac99509c97045327b4c40105b78ebccbf3ac04a6ea482e8989638a07" Oct 09 11:14:49 crc kubenswrapper[4923]: E1009 11:14:49.345602 4923 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 40s restarting failed container=manager pod=test-operator-controller-manager-776875b78c-5975g_openstack-operators(c703820f-a08c-4666-abc7-48954e40d927)\"" pod="openstack-operators/test-operator-controller-manager-776875b78c-5975g" podUID="c703820f-a08c-4666-abc7-48954e40d927" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071714462024454 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071714463017372 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071704010016500 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071704010015450 5ustar corecore